var/home/core/zuul-output/0000755000175000017500000000000015112022360014515 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015112031023015455 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004630564315112031013017673 0ustar rootrootNov 27 10:21:34 crc systemd[1]: Starting Kubernetes Kubelet... Nov 27 10:21:34 crc restorecon[4671]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:34 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 10:21:35 crc restorecon[4671]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 27 10:21:35 crc restorecon[4671]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 27 10:21:35 crc kubenswrapper[4806]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 10:21:35 crc kubenswrapper[4806]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 27 10:21:35 crc kubenswrapper[4806]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 10:21:35 crc kubenswrapper[4806]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 10:21:35 crc kubenswrapper[4806]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 27 10:21:35 crc kubenswrapper[4806]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.877685 4806 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.885951 4806 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886003 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886014 4806 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886026 4806 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886039 4806 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886061 4806 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886077 4806 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886087 4806 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886097 4806 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886112 4806 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886127 4806 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886142 4806 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886154 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886163 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886172 4806 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886182 4806 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886191 4806 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886200 4806 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886209 4806 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886218 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886271 4806 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886282 4806 feature_gate.go:330] unrecognized feature gate: Example Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886291 4806 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886299 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886308 4806 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886315 4806 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886324 4806 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886332 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886340 4806 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886347 4806 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886356 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886381 4806 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886393 4806 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886402 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886410 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886419 4806 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886427 4806 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886435 4806 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886443 4806 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886451 4806 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886459 4806 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886466 4806 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886475 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886483 4806 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886490 4806 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886498 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886507 4806 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886516 4806 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886523 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886532 4806 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886540 4806 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886548 4806 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886556 4806 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886563 4806 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886577 4806 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886586 4806 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886595 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886603 4806 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886611 4806 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886619 4806 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886627 4806 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886634 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886642 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886649 4806 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886657 4806 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886666 4806 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886673 4806 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886681 4806 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886689 4806 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886697 4806 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.886704 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.886863 4806 flags.go:64] FLAG: --address="0.0.0.0" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.886883 4806 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.886900 4806 flags.go:64] FLAG: --anonymous-auth="true" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.886912 4806 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.886924 4806 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.886934 4806 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.886948 4806 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.886960 4806 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.886969 4806 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.886979 4806 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.886989 4806 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887002 4806 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887011 4806 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887020 4806 flags.go:64] FLAG: --cgroup-root="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887029 4806 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887038 4806 flags.go:64] FLAG: --client-ca-file="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887047 4806 flags.go:64] FLAG: --cloud-config="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887055 4806 flags.go:64] FLAG: --cloud-provider="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887064 4806 flags.go:64] FLAG: --cluster-dns="[]" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887077 4806 flags.go:64] FLAG: --cluster-domain="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887086 4806 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887095 4806 flags.go:64] FLAG: --config-dir="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887104 4806 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887114 4806 flags.go:64] FLAG: --container-log-max-files="5" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887125 4806 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887134 4806 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887143 4806 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887152 4806 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887163 4806 flags.go:64] FLAG: --contention-profiling="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887175 4806 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887187 4806 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887199 4806 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887210 4806 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887279 4806 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887292 4806 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887304 4806 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887315 4806 flags.go:64] FLAG: --enable-load-reader="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887327 4806 flags.go:64] FLAG: --enable-server="true" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887339 4806 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887355 4806 flags.go:64] FLAG: --event-burst="100" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887368 4806 flags.go:64] FLAG: --event-qps="50" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887379 4806 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887391 4806 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887402 4806 flags.go:64] FLAG: --eviction-hard="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887418 4806 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887429 4806 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887441 4806 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887455 4806 flags.go:64] FLAG: --eviction-soft="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887466 4806 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887475 4806 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887484 4806 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887494 4806 flags.go:64] FLAG: --experimental-mounter-path="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887503 4806 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887511 4806 flags.go:64] FLAG: --fail-swap-on="true" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887520 4806 flags.go:64] FLAG: --feature-gates="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887532 4806 flags.go:64] FLAG: --file-check-frequency="20s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887541 4806 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887551 4806 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887559 4806 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887569 4806 flags.go:64] FLAG: --healthz-port="10248" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887578 4806 flags.go:64] FLAG: --help="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887587 4806 flags.go:64] FLAG: --hostname-override="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887595 4806 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887605 4806 flags.go:64] FLAG: --http-check-frequency="20s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887613 4806 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887622 4806 flags.go:64] FLAG: --image-credential-provider-config="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887631 4806 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887640 4806 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887648 4806 flags.go:64] FLAG: --image-service-endpoint="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887656 4806 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887667 4806 flags.go:64] FLAG: --kube-api-burst="100" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887676 4806 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887685 4806 flags.go:64] FLAG: --kube-api-qps="50" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887694 4806 flags.go:64] FLAG: --kube-reserved="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887703 4806 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887712 4806 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887721 4806 flags.go:64] FLAG: --kubelet-cgroups="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887730 4806 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887739 4806 flags.go:64] FLAG: --lock-file="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887747 4806 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887756 4806 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887765 4806 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887780 4806 flags.go:64] FLAG: --log-json-split-stream="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887790 4806 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887799 4806 flags.go:64] FLAG: --log-text-split-stream="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887807 4806 flags.go:64] FLAG: --logging-format="text" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887816 4806 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887825 4806 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887834 4806 flags.go:64] FLAG: --manifest-url="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887843 4806 flags.go:64] FLAG: --manifest-url-header="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887855 4806 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887864 4806 flags.go:64] FLAG: --max-open-files="1000000" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887875 4806 flags.go:64] FLAG: --max-pods="110" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887884 4806 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887893 4806 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887902 4806 flags.go:64] FLAG: --memory-manager-policy="None" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887911 4806 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887920 4806 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887929 4806 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887938 4806 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887961 4806 flags.go:64] FLAG: --node-status-max-images="50" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887971 4806 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887980 4806 flags.go:64] FLAG: --oom-score-adj="-999" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887989 4806 flags.go:64] FLAG: --pod-cidr="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.887997 4806 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888011 4806 flags.go:64] FLAG: --pod-manifest-path="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888020 4806 flags.go:64] FLAG: --pod-max-pids="-1" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888029 4806 flags.go:64] FLAG: --pods-per-core="0" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888038 4806 flags.go:64] FLAG: --port="10250" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888047 4806 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888056 4806 flags.go:64] FLAG: --provider-id="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888066 4806 flags.go:64] FLAG: --qos-reserved="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888075 4806 flags.go:64] FLAG: --read-only-port="10255" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888084 4806 flags.go:64] FLAG: --register-node="true" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888123 4806 flags.go:64] FLAG: --register-schedulable="true" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888133 4806 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888161 4806 flags.go:64] FLAG: --registry-burst="10" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888170 4806 flags.go:64] FLAG: --registry-qps="5" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888179 4806 flags.go:64] FLAG: --reserved-cpus="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888189 4806 flags.go:64] FLAG: --reserved-memory="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888200 4806 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888209 4806 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888219 4806 flags.go:64] FLAG: --rotate-certificates="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888262 4806 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888272 4806 flags.go:64] FLAG: --runonce="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888281 4806 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888290 4806 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888300 4806 flags.go:64] FLAG: --seccomp-default="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888309 4806 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888318 4806 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888329 4806 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888338 4806 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888348 4806 flags.go:64] FLAG: --storage-driver-password="root" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888356 4806 flags.go:64] FLAG: --storage-driver-secure="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888365 4806 flags.go:64] FLAG: --storage-driver-table="stats" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888376 4806 flags.go:64] FLAG: --storage-driver-user="root" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888385 4806 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888395 4806 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888405 4806 flags.go:64] FLAG: --system-cgroups="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888413 4806 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888428 4806 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888437 4806 flags.go:64] FLAG: --tls-cert-file="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888446 4806 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888457 4806 flags.go:64] FLAG: --tls-min-version="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888466 4806 flags.go:64] FLAG: --tls-private-key-file="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888476 4806 flags.go:64] FLAG: --topology-manager-policy="none" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888488 4806 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888499 4806 flags.go:64] FLAG: --topology-manager-scope="container" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888511 4806 flags.go:64] FLAG: --v="2" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888527 4806 flags.go:64] FLAG: --version="false" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888541 4806 flags.go:64] FLAG: --vmodule="" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888555 4806 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.888567 4806 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.888959 4806 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.888971 4806 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.888981 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.888989 4806 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.888997 4806 feature_gate.go:330] unrecognized feature gate: Example Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889006 4806 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889014 4806 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889023 4806 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889031 4806 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889038 4806 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889046 4806 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889054 4806 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889062 4806 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889069 4806 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889077 4806 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889085 4806 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889093 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889101 4806 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889108 4806 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889116 4806 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889124 4806 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889132 4806 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889140 4806 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889147 4806 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889155 4806 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889163 4806 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889171 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889178 4806 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889186 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889193 4806 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889201 4806 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889209 4806 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889216 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889224 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889259 4806 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889267 4806 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889275 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889283 4806 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889292 4806 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889300 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889310 4806 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889319 4806 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889328 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889336 4806 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889343 4806 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889351 4806 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889359 4806 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889367 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889375 4806 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889385 4806 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889393 4806 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889404 4806 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889415 4806 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889425 4806 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889435 4806 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889445 4806 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889455 4806 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889464 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889473 4806 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889481 4806 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889489 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889498 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889506 4806 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889514 4806 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889522 4806 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889530 4806 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889537 4806 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889545 4806 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889554 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889562 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.889569 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.889594 4806 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.898838 4806 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.898871 4806 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.898986 4806 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.898998 4806 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899007 4806 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899015 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899023 4806 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899031 4806 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899039 4806 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899046 4806 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899055 4806 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899063 4806 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899071 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899078 4806 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899086 4806 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899094 4806 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899102 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899109 4806 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899117 4806 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899125 4806 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899133 4806 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899141 4806 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899148 4806 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899157 4806 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899164 4806 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899173 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899181 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899189 4806 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899198 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899205 4806 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899213 4806 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899221 4806 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899260 4806 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899271 4806 feature_gate.go:330] unrecognized feature gate: Example Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899280 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899290 4806 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899304 4806 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899312 4806 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899320 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899328 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899336 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899344 4806 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899354 4806 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899366 4806 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899376 4806 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899385 4806 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899393 4806 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899402 4806 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899412 4806 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899421 4806 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899429 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899437 4806 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899445 4806 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899454 4806 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899462 4806 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899470 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899479 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899487 4806 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899495 4806 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899503 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899511 4806 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899519 4806 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899527 4806 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899535 4806 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899545 4806 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899555 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899566 4806 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899574 4806 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899583 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899591 4806 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899600 4806 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899607 4806 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899629 4806 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.899642 4806 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899868 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899881 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899890 4806 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899898 4806 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899907 4806 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899915 4806 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899923 4806 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899933 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899941 4806 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899950 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899958 4806 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899966 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899974 4806 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899982 4806 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899991 4806 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.899998 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900006 4806 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900014 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900023 4806 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900032 4806 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900039 4806 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900047 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900055 4806 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900063 4806 feature_gate.go:330] unrecognized feature gate: Example Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900071 4806 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900078 4806 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900112 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900120 4806 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900127 4806 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900135 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900144 4806 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900152 4806 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900160 4806 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900167 4806 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900176 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900185 4806 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900192 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900200 4806 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900208 4806 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900215 4806 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900225 4806 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900266 4806 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900280 4806 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900291 4806 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900299 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900308 4806 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900316 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900324 4806 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900332 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900340 4806 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900348 4806 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900358 4806 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900369 4806 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900377 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900386 4806 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900394 4806 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900403 4806 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900411 4806 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900420 4806 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900428 4806 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900437 4806 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900446 4806 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900455 4806 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900464 4806 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900472 4806 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900480 4806 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900491 4806 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900502 4806 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900510 4806 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900519 4806 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 27 10:21:35 crc kubenswrapper[4806]: W1127 10:21:35.900528 4806 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.900540 4806 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.901685 4806 server.go:940] "Client rotation is on, will bootstrap in background" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.907289 4806 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.907440 4806 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.910541 4806 server.go:997] "Starting client certificate rotation" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.910591 4806 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.910845 4806 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-04 14:32:27.232297145 +0000 UTC Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.911025 4806 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 916h10m51.321279393s for next certificate rotation Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.938648 4806 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.942096 4806 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 27 10:21:35 crc kubenswrapper[4806]: I1127 10:21:35.957650 4806 log.go:25] "Validated CRI v1 runtime API" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.002851 4806 log.go:25] "Validated CRI v1 image API" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.008334 4806 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.017207 4806 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-27-10-15-16-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.017255 4806 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.029813 4806 manager.go:217] Machine: {Timestamp:2025-11-27 10:21:36.027602408 +0000 UTC m=+0.614193192 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2799998 MemoryCapacity:25199476736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:0eb6d203-bd43-4b93-a96a-7190a2d0a1cc BootID:2d327c06-4eb5-4c2e-a11c-41971ef18290 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:61:9c:25 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:61:9c:25 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:68:15:58 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:f6:dc:0d Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:37:09:82 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:c6:5d:3e Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:a8:b7:b9 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:9e:ed:e2:c1:74:cd Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:46:26:f8:76:8d:a1 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199476736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.030029 4806 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.030146 4806 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.032173 4806 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.032750 4806 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.032813 4806 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.033144 4806 topology_manager.go:138] "Creating topology manager with none policy" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.033163 4806 container_manager_linux.go:303] "Creating device plugin manager" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.033753 4806 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.033805 4806 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.034048 4806 state_mem.go:36] "Initialized new in-memory state store" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.034178 4806 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.038358 4806 kubelet.go:418] "Attempting to sync node with API server" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.038397 4806 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.038432 4806 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.038452 4806 kubelet.go:324] "Adding apiserver pod source" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.038469 4806 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.044000 4806 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 27 10:21:36 crc kubenswrapper[4806]: W1127 10:21:36.044781 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.044883 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Nov 27 10:21:36 crc kubenswrapper[4806]: W1127 10:21:36.044855 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.044972 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.045999 4806 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.048526 4806 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.050260 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.050288 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.050299 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.050307 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.050321 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.050329 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.050339 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.050353 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.050363 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.050373 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.050404 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.050413 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.052274 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.052670 4806 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.052781 4806 server.go:1280] "Started kubelet" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.052959 4806 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.053463 4806 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 27 10:21:36 crc systemd[1]: Started Kubernetes Kubelet. Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.055782 4806 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.056739 4806 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.056779 4806 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.057028 4806 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 17:34:35.592786166 +0000 UTC Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.057112 4806 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 55h12m59.53568116s for next certificate rotation Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.057431 4806 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.057565 4806 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.058715 4806 server.go:460] "Adding debug handlers to kubelet server" Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.057485 4806 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.061013 4806 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 27 10:21:36 crc kubenswrapper[4806]: W1127 10:21:36.062031 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.063339 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.064722 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="200ms" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.065110 4806 factory.go:55] Registering systemd factory Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.065126 4806 factory.go:221] Registration of the systemd container factory successfully Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.065458 4806 factory.go:153] Registering CRI-O factory Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.065473 4806 factory.go:221] Registration of the crio container factory successfully Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.065558 4806 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.065589 4806 factory.go:103] Registering Raw factory Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.065619 4806 manager.go:1196] Started watching for new ooms in manager Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.066889 4806 manager.go:319] Starting recovery of all containers Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.069048 4806 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.246:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187bd5edba26dab0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 10:21:36.05275512 +0000 UTC m=+0.639345894,LastTimestamp:2025-11-27 10:21:36.05275512 +0000 UTC m=+0.639345894,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.079939 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080099 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080128 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080153 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080176 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080199 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080223 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080284 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080313 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080341 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080369 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080394 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080416 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080448 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080475 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080503 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080529 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080554 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080578 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080603 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080663 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080718 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080746 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080771 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080796 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080822 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080851 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080876 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080899 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080923 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080949 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.080978 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081006 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081030 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081055 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081080 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081103 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081129 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081154 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081178 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081201 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081225 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081283 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081308 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081333 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081356 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081382 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081408 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081432 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081457 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081480 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081506 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081538 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081564 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081594 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081621 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081647 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081673 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081696 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081719 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081744 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081768 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081792 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081817 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081842 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081867 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081891 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081918 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081944 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081969 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.081994 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.082020 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.082044 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.082068 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.082091 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.082116 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.082140 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.082162 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.082184 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.082209 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.082264 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.082308 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.082334 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087210 4806 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087324 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087458 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087502 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087529 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087558 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087589 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087616 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087642 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087672 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087700 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087728 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087757 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087785 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087812 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087837 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087862 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087885 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087908 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087936 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087962 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.087987 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088073 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088106 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088133 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088163 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088207 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088269 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088304 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088335 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088368 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088399 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088430 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088467 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088487 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088505 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088522 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088540 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088558 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088604 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088712 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088733 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088750 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088792 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.088812 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089550 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089573 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089623 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089647 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089665 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089682 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089698 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089713 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089731 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089749 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089764 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089779 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089793 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089809 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089825 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089841 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089856 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089900 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089916 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089930 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089945 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089960 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089975 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.089991 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090007 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090023 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090039 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090055 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090072 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090087 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090103 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090122 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090137 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090153 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090170 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090188 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090204 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090219 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090261 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090280 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090297 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090311 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090329 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090345 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090362 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090382 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090400 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090421 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090442 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090461 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090478 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090496 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090514 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090528 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090543 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090557 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090571 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090585 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090601 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090616 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090630 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090646 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090662 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090677 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090694 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090709 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090727 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090742 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090756 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090770 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090785 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090800 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090815 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090831 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090848 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090863 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090879 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090895 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090910 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090925 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090939 4806 reconstruct.go:97] "Volume reconstruction finished" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.090950 4806 reconciler.go:26] "Reconciler: start to sync state" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.106757 4806 manager.go:324] Recovery completed Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.112595 4806 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.115334 4806 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.115388 4806 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.115420 4806 kubelet.go:2335] "Starting kubelet main sync loop" Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.115535 4806 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.117067 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: W1127 10:21:36.117844 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.117904 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.118486 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.118510 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.118529 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.120405 4806 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.120428 4806 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.120448 4806 state_mem.go:36] "Initialized new in-memory state store" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.144572 4806 policy_none.go:49] "None policy: Start" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.145616 4806 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.145663 4806 state_mem.go:35] "Initializing new in-memory state store" Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.160978 4806 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.195443 4806 manager.go:334] "Starting Device Plugin manager" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.195495 4806 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.195507 4806 server.go:79] "Starting device plugin registration server" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.195851 4806 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.195862 4806 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.196190 4806 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.196272 4806 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.196281 4806 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.202396 4806 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.216564 4806 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.216648 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.217751 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.217778 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.217789 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.217905 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.218141 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.218191 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.218545 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.218572 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.218582 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.218738 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.218874 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.218911 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.219424 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.219444 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.219453 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.219611 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.219639 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.219649 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.219783 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.219907 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.219955 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.220536 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.220565 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.220574 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.221431 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.221459 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.221473 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.222657 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.222689 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.222701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.222811 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.222929 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.222959 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.223633 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.223645 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.223656 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.223669 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.223660 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.223705 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.223825 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.223850 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.224587 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.224613 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.224621 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.265356 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="400ms" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.292834 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.292914 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.292950 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.292981 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.293009 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.293042 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.293088 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.293125 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.293166 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.293211 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.293273 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.293332 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.293383 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.293418 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.293443 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.296405 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.297515 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.297557 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.297572 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.297599 4806 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.298074 4806 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.246:6443: connect: connection refused" node="crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395328 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395384 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395403 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395419 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395433 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395448 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395463 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395482 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395497 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395512 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395527 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395540 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395601 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395621 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395617 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395661 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395717 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395679 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395748 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395756 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395775 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395674 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395636 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395750 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395805 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395840 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395885 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395807 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395612 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.395907 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.499140 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.500629 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.500660 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.500672 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.500698 4806 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.501120 4806 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.246:6443: connect: connection refused" node="crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.558960 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.568432 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.597713 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: W1127 10:21:36.610782 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-ef86c01de3a3b1c5a2cb68b03387094a1d2f1ffa75468725e725951dfeac0e4a WatchSource:0}: Error finding container ef86c01de3a3b1c5a2cb68b03387094a1d2f1ffa75468725e725951dfeac0e4a: Status 404 returned error can't find the container with id ef86c01de3a3b1c5a2cb68b03387094a1d2f1ffa75468725e725951dfeac0e4a Nov 27 10:21:36 crc kubenswrapper[4806]: W1127 10:21:36.617175 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-328f5c3d4ed5e3c27a99bd2f0e47ab52d4e88916f3e6988023d14527cc3ef4ee WatchSource:0}: Error finding container 328f5c3d4ed5e3c27a99bd2f0e47ab52d4e88916f3e6988023d14527cc3ef4ee: Status 404 returned error can't find the container with id 328f5c3d4ed5e3c27a99bd2f0e47ab52d4e88916f3e6988023d14527cc3ef4ee Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.620889 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: W1127 10:21:36.628206 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-7de8a79636f38f7de1224c0f3772bf185b6dfd097f4ec2329a93f4aa629591e3 WatchSource:0}: Error finding container 7de8a79636f38f7de1224c0f3772bf185b6dfd097f4ec2329a93f4aa629591e3: Status 404 returned error can't find the container with id 7de8a79636f38f7de1224c0f3772bf185b6dfd097f4ec2329a93f4aa629591e3 Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.630288 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 10:21:36 crc kubenswrapper[4806]: W1127 10:21:36.659545 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-69caa46b4ae8ded010d606572ae01f758aef2b286c838ce91fcb6bbbf75f2e13 WatchSource:0}: Error finding container 69caa46b4ae8ded010d606572ae01f758aef2b286c838ce91fcb6bbbf75f2e13: Status 404 returned error can't find the container with id 69caa46b4ae8ded010d606572ae01f758aef2b286c838ce91fcb6bbbf75f2e13 Nov 27 10:21:36 crc kubenswrapper[4806]: W1127 10:21:36.660344 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-56bfbe230ccbf0eba64b437ce82911438f920bafaa387cd30e85a2527280fb70 WatchSource:0}: Error finding container 56bfbe230ccbf0eba64b437ce82911438f920bafaa387cd30e85a2527280fb70: Status 404 returned error can't find the container with id 56bfbe230ccbf0eba64b437ce82911438f920bafaa387cd30e85a2527280fb70 Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.666602 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="800ms" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.902137 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.903040 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.903069 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.903077 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:36 crc kubenswrapper[4806]: I1127 10:21:36.903097 4806 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.903516 4806 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.246:6443: connect: connection refused" node="crc" Nov 27 10:21:36 crc kubenswrapper[4806]: W1127 10:21:36.927032 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.927130 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Nov 27 10:21:36 crc kubenswrapper[4806]: W1127 10:21:36.985565 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:36 crc kubenswrapper[4806]: E1127 10:21:36.985667 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Nov 27 10:21:37 crc kubenswrapper[4806]: I1127 10:21:37.053673 4806 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:37 crc kubenswrapper[4806]: I1127 10:21:37.120438 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ef86c01de3a3b1c5a2cb68b03387094a1d2f1ffa75468725e725951dfeac0e4a"} Nov 27 10:21:37 crc kubenswrapper[4806]: I1127 10:21:37.121398 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"56bfbe230ccbf0eba64b437ce82911438f920bafaa387cd30e85a2527280fb70"} Nov 27 10:21:37 crc kubenswrapper[4806]: I1127 10:21:37.122176 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"69caa46b4ae8ded010d606572ae01f758aef2b286c838ce91fcb6bbbf75f2e13"} Nov 27 10:21:37 crc kubenswrapper[4806]: I1127 10:21:37.122924 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7de8a79636f38f7de1224c0f3772bf185b6dfd097f4ec2329a93f4aa629591e3"} Nov 27 10:21:37 crc kubenswrapper[4806]: I1127 10:21:37.123657 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"328f5c3d4ed5e3c27a99bd2f0e47ab52d4e88916f3e6988023d14527cc3ef4ee"} Nov 27 10:21:37 crc kubenswrapper[4806]: W1127 10:21:37.357643 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:37 crc kubenswrapper[4806]: E1127 10:21:37.358203 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Nov 27 10:21:37 crc kubenswrapper[4806]: E1127 10:21:37.468285 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="1.6s" Nov 27 10:21:37 crc kubenswrapper[4806]: W1127 10:21:37.625359 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:37 crc kubenswrapper[4806]: E1127 10:21:37.625492 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Nov 27 10:21:37 crc kubenswrapper[4806]: I1127 10:21:37.704301 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:37 crc kubenswrapper[4806]: I1127 10:21:37.707174 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:37 crc kubenswrapper[4806]: I1127 10:21:37.707266 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:37 crc kubenswrapper[4806]: I1127 10:21:37.707285 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:37 crc kubenswrapper[4806]: I1127 10:21:37.707333 4806 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 10:21:37 crc kubenswrapper[4806]: E1127 10:21:37.708719 4806 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.246:6443: connect: connection refused" node="crc" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.053916 4806 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.127641 4806 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4" exitCode=0 Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.127708 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4"} Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.127731 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.128824 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.128864 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.128875 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.130010 4806 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71" exitCode=0 Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.130055 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71"} Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.130157 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.131371 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.131399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.131410 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.132213 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28" exitCode=0 Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.132270 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28"} Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.132302 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.133657 4806 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee" exitCode=0 Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.133712 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee"} Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.133719 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.133746 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.133758 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.133798 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.134637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.134664 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.134675 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.136912 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9"} Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.136953 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3"} Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.136966 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01"} Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.136977 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8"} Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.137067 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.137881 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.137903 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.137913 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.138920 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.139848 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.139866 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.139885 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.362737 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:38 crc kubenswrapper[4806]: I1127 10:21:38.600565 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:38 crc kubenswrapper[4806]: W1127 10:21:38.915706 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:38 crc kubenswrapper[4806]: E1127 10:21:38.915770 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.053476 4806 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:39 crc kubenswrapper[4806]: E1127 10:21:39.069772 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="3.2s" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.141875 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"74fe72cadf6271151c95f0efb2cd1e154c78d1398d9f104c57acc255ecc99738"} Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.141936 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7b2d5c2fdd37029ccfa325eb3a2f2fa6d1bcb65b3847dd4997e2cc12798603c2"} Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.141947 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"dc6784c0e140b36ee0c0a9135631f867702e99bd54be29abcea1bd7d97f5c533"} Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.142070 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.143025 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.143056 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.143067 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.146326 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280"} Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.146352 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310"} Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.146362 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12"} Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.146374 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61"} Nov 27 10:21:39 crc kubenswrapper[4806]: W1127 10:21:39.147636 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:39 crc kubenswrapper[4806]: E1127 10:21:39.147705 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.149546 4806 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70" exitCode=0 Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.149673 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70"} Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.149722 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.151129 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.151173 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.151188 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.155315 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.155390 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.156059 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1a7df8a04a47878ce0bd708ebe5559eb0a1e784e3c53749d1790741e8713046c"} Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.156685 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.156719 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.156733 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.157362 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.157399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.157416 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:39 crc kubenswrapper[4806]: W1127 10:21:39.175196 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:39 crc kubenswrapper[4806]: E1127 10:21:39.175413 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.309167 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.312973 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.313018 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.313031 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:39 crc kubenswrapper[4806]: I1127 10:21:39.313059 4806 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 10:21:39 crc kubenswrapper[4806]: E1127 10:21:39.314315 4806 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.246:6443: connect: connection refused" node="crc" Nov 27 10:21:39 crc kubenswrapper[4806]: W1127 10:21:39.552084 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.246:6443: connect: connection refused Nov 27 10:21:39 crc kubenswrapper[4806]: E1127 10:21:39.552268 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.246:6443: connect: connection refused" logger="UnhandledError" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.168661 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127"} Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.168825 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.169663 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.169696 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.169705 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.172588 4806 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4" exitCode=0 Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.172676 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.172706 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.173123 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.173407 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4"} Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.173478 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.173830 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.174521 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.174542 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.174550 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.174989 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.175007 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.175015 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.176044 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.176076 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.176087 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.176173 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.176222 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.176257 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.181918 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:40 crc kubenswrapper[4806]: I1127 10:21:40.945627 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:41 crc kubenswrapper[4806]: I1127 10:21:41.177963 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 10:21:41 crc kubenswrapper[4806]: I1127 10:21:41.178014 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:41 crc kubenswrapper[4806]: I1127 10:21:41.178436 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea"} Nov 27 10:21:41 crc kubenswrapper[4806]: I1127 10:21:41.178505 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3"} Nov 27 10:21:41 crc kubenswrapper[4806]: I1127 10:21:41.178521 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45"} Nov 27 10:21:41 crc kubenswrapper[4806]: I1127 10:21:41.178529 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707"} Nov 27 10:21:41 crc kubenswrapper[4806]: I1127 10:21:41.178932 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:41 crc kubenswrapper[4806]: I1127 10:21:41.178966 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:41 crc kubenswrapper[4806]: I1127 10:21:41.178977 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:41 crc kubenswrapper[4806]: I1127 10:21:41.362992 4806 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 27 10:21:41 crc kubenswrapper[4806]: I1127 10:21:41.363086 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.073635 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.074163 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.075772 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.075817 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.075831 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.188210 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de"} Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.188326 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.189848 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.188396 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.191490 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.191552 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.191705 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.191727 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.191667 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.192152 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.509214 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.509810 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.511413 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.511484 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.511498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.514483 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.515861 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.515906 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.515920 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.516000 4806 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.521716 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:42 crc kubenswrapper[4806]: I1127 10:21:42.771394 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 27 10:21:43 crc kubenswrapper[4806]: I1127 10:21:43.190944 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:43 crc kubenswrapper[4806]: I1127 10:21:43.191519 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:43 crc kubenswrapper[4806]: I1127 10:21:43.194200 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:43 crc kubenswrapper[4806]: I1127 10:21:43.194250 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:43 crc kubenswrapper[4806]: I1127 10:21:43.194264 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:43 crc kubenswrapper[4806]: I1127 10:21:43.194358 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:43 crc kubenswrapper[4806]: I1127 10:21:43.194392 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:43 crc kubenswrapper[4806]: I1127 10:21:43.194404 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:43 crc kubenswrapper[4806]: I1127 10:21:43.694465 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:43 crc kubenswrapper[4806]: I1127 10:21:43.694752 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:43 crc kubenswrapper[4806]: I1127 10:21:43.696390 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:43 crc kubenswrapper[4806]: I1127 10:21:43.696457 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:43 crc kubenswrapper[4806]: I1127 10:21:43.696479 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:44 crc kubenswrapper[4806]: I1127 10:21:44.193631 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:44 crc kubenswrapper[4806]: I1127 10:21:44.194846 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:44 crc kubenswrapper[4806]: I1127 10:21:44.194906 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:44 crc kubenswrapper[4806]: I1127 10:21:44.194915 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:45 crc kubenswrapper[4806]: I1127 10:21:45.641297 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 10:21:45 crc kubenswrapper[4806]: I1127 10:21:45.641955 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:45 crc kubenswrapper[4806]: I1127 10:21:45.643196 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:45 crc kubenswrapper[4806]: I1127 10:21:45.643224 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:45 crc kubenswrapper[4806]: I1127 10:21:45.643273 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:46 crc kubenswrapper[4806]: E1127 10:21:46.202830 4806 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 27 10:21:48 crc kubenswrapper[4806]: I1127 10:21:48.423582 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 27 10:21:48 crc kubenswrapper[4806]: I1127 10:21:48.424300 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:48 crc kubenswrapper[4806]: I1127 10:21:48.425744 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:48 crc kubenswrapper[4806]: I1127 10:21:48.425798 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:48 crc kubenswrapper[4806]: I1127 10:21:48.425855 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.054581 4806 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.211948 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.214361 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127" exitCode=255 Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.214399 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127"} Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.214531 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.215346 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.215388 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.215402 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.216027 4806 scope.go:117] "RemoveContainer" containerID="355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127" Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.438492 4806 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.438545 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.443304 4806 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.443356 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.953633 4806 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]log ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]etcd ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/openshift.io-api-request-count-filter ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/openshift.io-startkubeinformers ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/generic-apiserver-start-informers ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/priority-and-fairness-config-consumer ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/priority-and-fairness-filter ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/start-apiextensions-informers ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/start-apiextensions-controllers ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/crd-informer-synced ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/start-system-namespaces-controller ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/start-cluster-authentication-info-controller ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/start-legacy-token-tracking-controller ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/start-service-ip-repair-controllers ok Nov 27 10:21:50 crc kubenswrapper[4806]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Nov 27 10:21:50 crc kubenswrapper[4806]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/priority-and-fairness-config-producer ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/bootstrap-controller ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/start-kube-aggregator-informers ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/apiservice-status-local-available-controller ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/apiservice-status-remote-available-controller ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/apiservice-registration-controller ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/apiservice-wait-for-first-sync ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/apiservice-discovery-controller ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/kube-apiserver-autoregistration ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]autoregister-completion ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/apiservice-openapi-controller ok Nov 27 10:21:50 crc kubenswrapper[4806]: [+]poststarthook/apiservice-openapiv3-controller ok Nov 27 10:21:50 crc kubenswrapper[4806]: livez check failed Nov 27 10:21:50 crc kubenswrapper[4806]: I1127 10:21:50.953730 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:21:51 crc kubenswrapper[4806]: I1127 10:21:51.219703 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 27 10:21:51 crc kubenswrapper[4806]: I1127 10:21:51.222314 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411"} Nov 27 10:21:51 crc kubenswrapper[4806]: I1127 10:21:51.222475 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:51 crc kubenswrapper[4806]: I1127 10:21:51.223385 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:51 crc kubenswrapper[4806]: I1127 10:21:51.223414 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:51 crc kubenswrapper[4806]: I1127 10:21:51.223425 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:51 crc kubenswrapper[4806]: I1127 10:21:51.363127 4806 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 27 10:21:51 crc kubenswrapper[4806]: I1127 10:21:51.363297 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 27 10:21:52 crc kubenswrapper[4806]: I1127 10:21:52.078540 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:52 crc kubenswrapper[4806]: I1127 10:21:52.078654 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:52 crc kubenswrapper[4806]: I1127 10:21:52.079631 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:52 crc kubenswrapper[4806]: I1127 10:21:52.079659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:52 crc kubenswrapper[4806]: I1127 10:21:52.079668 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:53 crc kubenswrapper[4806]: I1127 10:21:53.694990 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:53 crc kubenswrapper[4806]: I1127 10:21:53.695149 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:21:53 crc kubenswrapper[4806]: I1127 10:21:53.696280 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:21:53 crc kubenswrapper[4806]: I1127 10:21:53.696313 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:21:53 crc kubenswrapper[4806]: I1127 10:21:53.696326 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:21:55 crc kubenswrapper[4806]: E1127 10:21:55.458658 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 27 10:21:55 crc kubenswrapper[4806]: I1127 10:21:55.463387 4806 trace.go:236] Trace[1478003453]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 10:21:44.879) (total time: 10584ms): Nov 27 10:21:55 crc kubenswrapper[4806]: Trace[1478003453]: ---"Objects listed" error: 10584ms (10:21:55.463) Nov 27 10:21:55 crc kubenswrapper[4806]: Trace[1478003453]: [10.584162602s] [10.584162602s] END Nov 27 10:21:55 crc kubenswrapper[4806]: I1127 10:21:55.463439 4806 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 27 10:21:55 crc kubenswrapper[4806]: I1127 10:21:55.467885 4806 trace.go:236] Trace[1060750489]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 10:21:45.286) (total time: 10180ms): Nov 27 10:21:55 crc kubenswrapper[4806]: Trace[1060750489]: ---"Objects listed" error: 10180ms (10:21:55.467) Nov 27 10:21:55 crc kubenswrapper[4806]: Trace[1060750489]: [10.180817453s] [10.180817453s] END Nov 27 10:21:55 crc kubenswrapper[4806]: I1127 10:21:55.467938 4806 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 27 10:21:55 crc kubenswrapper[4806]: I1127 10:21:55.468160 4806 trace.go:236] Trace[1939222430]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 10:21:42.627) (total time: 12840ms): Nov 27 10:21:55 crc kubenswrapper[4806]: Trace[1939222430]: ---"Objects listed" error: 12840ms (10:21:55.467) Nov 27 10:21:55 crc kubenswrapper[4806]: Trace[1939222430]: [12.840755819s] [12.840755819s] END Nov 27 10:21:55 crc kubenswrapper[4806]: I1127 10:21:55.468204 4806 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 27 10:21:55 crc kubenswrapper[4806]: I1127 10:21:55.469298 4806 trace.go:236] Trace[1200289848]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Nov-2025 10:21:44.446) (total time: 11022ms): Nov 27 10:21:55 crc kubenswrapper[4806]: Trace[1200289848]: ---"Objects listed" error: 11022ms (10:21:55.468) Nov 27 10:21:55 crc kubenswrapper[4806]: Trace[1200289848]: [11.022463657s] [11.022463657s] END Nov 27 10:21:55 crc kubenswrapper[4806]: I1127 10:21:55.469676 4806 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 27 10:21:55 crc kubenswrapper[4806]: I1127 10:21:55.472349 4806 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 27 10:21:55 crc kubenswrapper[4806]: E1127 10:21:55.472584 4806 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 27 10:21:55 crc kubenswrapper[4806]: I1127 10:21:55.952471 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:55 crc kubenswrapper[4806]: I1127 10:21:55.963977 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.050385 4806 apiserver.go:52] "Watching apiserver" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.053071 4806 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.053490 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.053943 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.054004 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.054108 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.054213 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.054373 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.054295 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.054369 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.054707 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.054795 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.058567 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.060032 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.065624 4806 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.065797 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.066405 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.066709 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.066891 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.067721 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.068590 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.068631 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.076917 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.076950 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.076971 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.076993 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077007 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077023 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077040 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077055 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077069 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077085 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077101 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077119 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077136 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077152 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077170 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077186 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077202 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077221 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077252 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077269 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077285 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077301 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077318 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077334 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077352 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077368 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077384 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077447 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077486 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077506 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077522 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077542 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077559 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077577 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077596 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077613 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077629 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077662 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077679 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077696 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077714 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077730 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077768 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077785 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077801 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077856 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077874 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077892 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077923 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077942 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077962 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077979 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.077994 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078025 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078040 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078055 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078071 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078086 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078101 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078123 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078141 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078157 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078172 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078187 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078205 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078219 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078249 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078268 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078286 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078301 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078472 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078655 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078676 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078692 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078721 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078741 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078757 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078863 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078881 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078897 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078914 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078928 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078948 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078965 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078982 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.078998 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079014 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079028 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079043 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079058 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079074 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079091 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079109 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079127 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079145 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079161 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079153 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079177 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079272 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079299 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079324 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079345 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079362 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079382 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079401 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079418 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079436 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079450 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079461 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079489 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079507 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079524 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079540 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079559 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079575 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079595 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079611 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079632 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079648 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079665 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079685 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079705 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079730 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079758 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079785 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079813 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079839 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079866 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079892 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079916 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079942 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079967 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.079990 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080017 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080040 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080065 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080088 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080114 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080139 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080141 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080165 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080195 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080219 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080264 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080290 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080319 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080348 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080374 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080399 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080429 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080457 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080482 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080503 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080519 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080535 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080551 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080561 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080567 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080599 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080622 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080645 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080664 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080683 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080703 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080770 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080791 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080812 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080818 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080832 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080851 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080870 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080890 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080908 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080927 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080947 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080965 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080968 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.080985 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081003 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081022 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081039 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081057 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081073 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081089 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081099 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081106 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081139 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081160 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081179 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081198 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081219 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081278 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081302 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081325 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081347 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081371 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081392 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081409 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081425 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081442 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081461 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081466 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081514 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081574 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081614 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081725 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081777 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081808 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081834 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081863 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081891 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081916 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081944 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.081969 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.082077 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.082167 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.082197 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.082324 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.082344 4806 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.082361 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.082375 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.082390 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.082405 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.082419 4806 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.083165 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.083193 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.083517 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.083575 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.083762 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.083893 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.083930 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.084011 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.084390 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.084662 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.084546 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.084838 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.085015 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.085016 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.085065 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.085576 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.085645 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:21:56.585621447 +0000 UTC m=+21.172212411 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.085665 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.085764 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.085783 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.086112 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.086161 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.086668 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.086795 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.087004 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.087441 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.088799 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.089307 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.089498 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.089835 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.090001 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.090643 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.090704 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.090875 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:21:56.590842686 +0000 UTC m=+21.177433460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.091058 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.091827 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.091981 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.092206 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.092598 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.092893 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.092900 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.093157 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.093422 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.093524 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.093615 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.093844 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.093947 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.094010 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.094072 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.094529 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.094704 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.094814 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.095091 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.095118 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.099577 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.099742 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.099965 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.100049 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.100277 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.100478 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.100806 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.101026 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.101703 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.102210 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.102536 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.102772 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.102877 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.103148 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.103281 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.103494 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.103750 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.104501 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.106300 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.106792 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.107002 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.107893 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.107967 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.108055 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.108067 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.108171 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.108903 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.109342 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.109731 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.109925 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.109966 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.110202 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.110275 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.110508 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.110529 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.110748 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.110955 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.111376 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.111414 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.111852 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.111856 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.112105 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.112362 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.112444 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.112690 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.113122 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.113318 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.113445 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.113580 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.113861 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.114084 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.114221 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.115147 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.115499 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.115522 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.115569 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.115683 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.115727 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.115743 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.116060 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.116083 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.116224 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.116491 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.116702 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.117308 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.111531 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.116944 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.117925 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.118494 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.123480 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.118596 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.118815 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.118847 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.119094 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.119279 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.119467 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.119469 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.119795 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.119802 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.120263 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.120287 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.120428 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.120481 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.120677 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.120914 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.121049 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.121206 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.121470 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.121528 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.121714 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.121804 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.121830 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.122557 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.122674 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.123890 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.123269 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.124010 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.124051 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.124564 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.124568 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.124905 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.125294 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.125684 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.126228 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.126488 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.126767 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.127591 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.127634 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.127713 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.128278 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.128566 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.128701 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.128830 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.129034 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.129553 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.129908 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.130089 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.130251 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.130350 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.130575 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.131165 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.131402 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.132133 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.132290 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:21:56.632265699 +0000 UTC m=+21.218856463 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.132523 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.132755 4806 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.132880 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.132952 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.133601 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.133754 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.133290 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.134962 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.135469 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.139523 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.140048 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.141780 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.147383 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.147570 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.147565 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.147739 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.155987 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.169034 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.169626 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.169650 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.169664 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.169726 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 10:21:56.66970606 +0000 UTC m=+21.256296824 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.175520 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.175557 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.175572 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.175629 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 10:21:56.675608444 +0000 UTC m=+21.262199208 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.176649 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.176769 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184643 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184694 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184744 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184754 4806 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184763 4806 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184771 4806 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184781 4806 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184789 4806 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184798 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184807 4806 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184817 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184825 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184833 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184842 4806 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184850 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184858 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184869 4806 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184878 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184886 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184895 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184903 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184904 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184913 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.184996 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185010 4806 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185022 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185036 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185051 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185063 4806 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185075 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185087 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185099 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185113 4806 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185135 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185148 4806 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185158 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185167 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185176 4806 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185184 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185193 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185202 4806 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185211 4806 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185221 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185245 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185253 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185262 4806 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185270 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185278 4806 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185286 4806 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185295 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185303 4806 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185311 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185319 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185327 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185335 4806 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185343 4806 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185351 4806 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185359 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185366 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185374 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185382 4806 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185390 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185398 4806 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185406 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185414 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185421 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185428 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185436 4806 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185445 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185454 4806 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185461 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185469 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185481 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185489 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185497 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185504 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185512 4806 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185520 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185528 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185536 4806 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185544 4806 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185552 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185561 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185606 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185622 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185701 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185758 4806 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185777 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185790 4806 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185804 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185816 4806 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185829 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185843 4806 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185914 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185926 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185939 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185953 4806 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185965 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185976 4806 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185987 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.185998 4806 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186017 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186029 4806 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186041 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186053 4806 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186065 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186077 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186089 4806 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186101 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186113 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186124 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186137 4806 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186149 4806 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186160 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186175 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186188 4806 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186199 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186209 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186219 4806 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186250 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186262 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186272 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186282 4806 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186293 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186304 4806 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186315 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186325 4806 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186336 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186347 4806 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186358 4806 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186369 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186379 4806 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186390 4806 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186421 4806 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186431 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186443 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186453 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186463 4806 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186474 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186484 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186495 4806 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186507 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186517 4806 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186528 4806 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186538 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186549 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186563 4806 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186574 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186586 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186598 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186608 4806 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186619 4806 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186629 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186640 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186651 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186662 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186673 4806 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186684 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186694 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186704 4806 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186714 4806 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186727 4806 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186738 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186749 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186761 4806 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186772 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186784 4806 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186794 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186805 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186815 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186825 4806 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186836 4806 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186851 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186861 4806 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186871 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186882 4806 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186893 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186904 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186916 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186927 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186937 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186947 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186957 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186969 4806 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186979 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.186990 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.187005 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.187017 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.187027 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.202260 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.202908 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.205865 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.206826 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.215647 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.218187 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.221907 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.233697 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.234934 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.236295 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.237213 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.240106 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.241052 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.242060 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.247813 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.248541 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.250220 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.250810 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.251655 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.252104 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.252144 4806 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.254869 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.255757 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.259984 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.266174 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.268006 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.268840 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.270497 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.271331 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.272322 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.273156 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.274938 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.276299 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.277515 4806 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.277783 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.279576 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.281422 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.282837 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.283893 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.287485 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.287589 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.288077 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.288104 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.288117 4806 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.289624 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.292121 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.293142 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.294591 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.295338 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.296696 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.297484 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.298626 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.299307 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.299971 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.300646 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.301527 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.303127 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.303836 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.305326 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.305978 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.306906 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.308110 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.308799 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.312841 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.322360 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.332814 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.341998 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.351746 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.365333 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.375351 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.376580 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.385763 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: W1127 10:21:56.387112 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-7d5da880a3aa93d3032d7f82c53ad026d641c5003fdfe56d188d22b4c3f83dc8 WatchSource:0}: Error finding container 7d5da880a3aa93d3032d7f82c53ad026d641c5003fdfe56d188d22b4c3f83dc8: Status 404 returned error can't find the container with id 7d5da880a3aa93d3032d7f82c53ad026d641c5003fdfe56d188d22b4c3f83dc8 Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.391080 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.395658 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.421455 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.421709 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 27 10:21:56 crc kubenswrapper[4806]: W1127 10:21:56.435158 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-cf9ee280b85444d2d146d7f275e6ff3edc3da07943135227aab73e6fe48ca8ba WatchSource:0}: Error finding container cf9ee280b85444d2d146d7f275e6ff3edc3da07943135227aab73e6fe48ca8ba: Status 404 returned error can't find the container with id cf9ee280b85444d2d146d7f275e6ff3edc3da07943135227aab73e6fe48ca8ba Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.589972 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.590173 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.590601 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:21:57.590580587 +0000 UTC m=+22.177171361 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.691160 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.691242 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.691265 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:21:56 crc kubenswrapper[4806]: I1127 10:21:56.691282 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.691382 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.691398 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.691408 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.691453 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 10:21:57.691427022 +0000 UTC m=+22.278017776 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.691543 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:21:57.691535224 +0000 UTC m=+22.278125988 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.691594 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.691605 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.691619 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.691638 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 10:21:57.691632786 +0000 UTC m=+22.278223550 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.691669 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:21:56 crc kubenswrapper[4806]: E1127 10:21:56.691688 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:21:57.691681887 +0000 UTC m=+22.278272651 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.245674 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2"} Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.245722 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"cf9ee280b85444d2d146d7f275e6ff3edc3da07943135227aab73e6fe48ca8ba"} Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.247682 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53"} Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.247718 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539"} Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.247730 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cd15864efa1e99609480b937a8ad2cbd6a1436ad03b4c54cb18c5b110be00788"} Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.249769 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7d5da880a3aa93d3032d7f82c53ad026d641c5003fdfe56d188d22b4c3f83dc8"} Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.271932 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.324706 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.352591 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.404323 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.444367 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.469167 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.495250 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.515467 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.532207 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.549941 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.572308 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.592318 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.598515 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:21:57 crc kubenswrapper[4806]: E1127 10:21:57.598721 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:21:57 crc kubenswrapper[4806]: E1127 10:21:57.598812 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:21:59.598793719 +0000 UTC m=+24.185384473 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.620937 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.682118 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.699389 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.699495 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.699521 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:21:57 crc kubenswrapper[4806]: E1127 10:21:57.699587 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:21:59.699560122 +0000 UTC m=+24.286150876 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:21:57 crc kubenswrapper[4806]: E1127 10:21:57.699608 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:21:57 crc kubenswrapper[4806]: E1127 10:21:57.699610 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:21:57 crc kubenswrapper[4806]: E1127 10:21:57.699623 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:21:57 crc kubenswrapper[4806]: E1127 10:21:57.699636 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:21:57 crc kubenswrapper[4806]: E1127 10:21:57.699639 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:57 crc kubenswrapper[4806]: E1127 10:21:57.699651 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.699645 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:21:57 crc kubenswrapper[4806]: E1127 10:21:57.699685 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 10:21:59.699672694 +0000 UTC m=+24.286263458 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:57 crc kubenswrapper[4806]: E1127 10:21:57.699699 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 10:21:59.699693805 +0000 UTC m=+24.286284559 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:57 crc kubenswrapper[4806]: E1127 10:21:57.699699 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:21:57 crc kubenswrapper[4806]: E1127 10:21:57.699741 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:21:59.699727006 +0000 UTC m=+24.286317770 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.700639 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-68rwq"] Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.700932 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-68rwq" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.703543 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.703822 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.704051 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.724498 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.737838 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.749871 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.764928 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.775498 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.787110 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.798848 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.800105 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9fld\" (UniqueName: \"kubernetes.io/projected/3dfed96b-3244-4631-b080-cf112ad196dc-kube-api-access-c9fld\") pod \"node-resolver-68rwq\" (UID: \"3dfed96b-3244-4631-b080-cf112ad196dc\") " pod="openshift-dns/node-resolver-68rwq" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.800181 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3dfed96b-3244-4631-b080-cf112ad196dc-hosts-file\") pod \"node-resolver-68rwq\" (UID: \"3dfed96b-3244-4631-b080-cf112ad196dc\") " pod="openshift-dns/node-resolver-68rwq" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.810702 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.901376 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3dfed96b-3244-4631-b080-cf112ad196dc-hosts-file\") pod \"node-resolver-68rwq\" (UID: \"3dfed96b-3244-4631-b080-cf112ad196dc\") " pod="openshift-dns/node-resolver-68rwq" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.901479 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9fld\" (UniqueName: \"kubernetes.io/projected/3dfed96b-3244-4631-b080-cf112ad196dc-kube-api-access-c9fld\") pod \"node-resolver-68rwq\" (UID: \"3dfed96b-3244-4631-b080-cf112ad196dc\") " pod="openshift-dns/node-resolver-68rwq" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.901607 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3dfed96b-3244-4631-b080-cf112ad196dc-hosts-file\") pod \"node-resolver-68rwq\" (UID: \"3dfed96b-3244-4631-b080-cf112ad196dc\") " pod="openshift-dns/node-resolver-68rwq" Nov 27 10:21:57 crc kubenswrapper[4806]: I1127 10:21:57.926213 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9fld\" (UniqueName: \"kubernetes.io/projected/3dfed96b-3244-4631-b080-cf112ad196dc-kube-api-access-c9fld\") pod \"node-resolver-68rwq\" (UID: \"3dfed96b-3244-4631-b080-cf112ad196dc\") " pod="openshift-dns/node-resolver-68rwq" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.012583 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-68rwq" Nov 27 10:21:58 crc kubenswrapper[4806]: W1127 10:21:58.024427 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dfed96b_3244_4631_b080_cf112ad196dc.slice/crio-7377c9d3f61c1e9a8124b9f636a8dd84c98784e1e417d9bca65c10418f7e06d4 WatchSource:0}: Error finding container 7377c9d3f61c1e9a8124b9f636a8dd84c98784e1e417d9bca65c10418f7e06d4: Status 404 returned error can't find the container with id 7377c9d3f61c1e9a8124b9f636a8dd84c98784e1e417d9bca65c10418f7e06d4 Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.103729 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-7mq9p"] Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.104010 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-44ndp"] Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.104163 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.104871 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.105784 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rsgmm"] Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.106426 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.111332 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-5cx6g"] Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.111708 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.117647 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:21:58 crc kubenswrapper[4806]: E1127 10:21:58.117769 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.117824 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:21:58 crc kubenswrapper[4806]: E1127 10:21:58.117861 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.117966 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.118348 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.118501 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.118609 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.118716 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.118808 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.118878 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.118967 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 27 10:21:58 crc kubenswrapper[4806]: E1127 10:21:58.118994 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.119165 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.119308 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.119426 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.119546 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.121262 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.121526 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.121722 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.122663 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.123732 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.124513 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.130116 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.130637 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.137567 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.139964 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.144819 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.163303 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.182936 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.195547 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.203905 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-run-netns\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.203934 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-systemd\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.203948 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-ovn\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.203962 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-cni-bin\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.203977 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/39930cf4-5f3c-42cc-9384-d366bb1d6a78-cni-binary-copy\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.203992 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-var-lib-cni-multus\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204006 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-cni-binary-copy\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204019 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovnkube-config\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204034 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-system-cni-dir\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204048 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-multus-conf-dir\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204061 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-hostroot\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204077 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-run-multus-certs\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204094 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-var-lib-kubelet\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204108 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-slash\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204122 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-var-lib-cni-bin\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204136 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/39930cf4-5f3c-42cc-9384-d366bb1d6a78-multus-daemon-config\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204149 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204164 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ea0cad1-8e8b-458c-a943-7191b36e35a3-mcd-auth-proxy-config\") pod \"machine-config-daemon-7mq9p\" (UID: \"4ea0cad1-8e8b-458c-a943-7191b36e35a3\") " pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204180 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204196 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-log-socket\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204212 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cknt9\" (UniqueName: \"kubernetes.io/projected/4ea0cad1-8e8b-458c-a943-7191b36e35a3-kube-api-access-cknt9\") pod \"machine-config-daemon-7mq9p\" (UID: \"4ea0cad1-8e8b-458c-a943-7191b36e35a3\") " pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204239 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-kubelet\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204263 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-multus-cni-dir\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204280 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ea0cad1-8e8b-458c-a943-7191b36e35a3-proxy-tls\") pod \"machine-config-daemon-7mq9p\" (UID: \"4ea0cad1-8e8b-458c-a943-7191b36e35a3\") " pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204294 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-os-release\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204308 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204323 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65952\" (UniqueName: \"kubernetes.io/projected/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-kube-api-access-65952\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204342 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-run-k8s-cni-cncf-io\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204356 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njxg4\" (UniqueName: \"kubernetes.io/projected/39930cf4-5f3c-42cc-9384-d366bb1d6a78-kube-api-access-njxg4\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204370 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-cnibin\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204385 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4ea0cad1-8e8b-458c-a943-7191b36e35a3-rootfs\") pod \"machine-config-daemon-7mq9p\" (UID: \"4ea0cad1-8e8b-458c-a943-7191b36e35a3\") " pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204400 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-node-log\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204414 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-cnibin\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204428 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-env-overrides\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204442 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovn-node-metrics-cert\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204457 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovnkube-script-lib\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204470 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-os-release\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204493 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-systemd-units\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204507 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-cni-netd\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204529 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czmdn\" (UniqueName: \"kubernetes.io/projected/e8ed6bbc-965a-412c-a3ee-4763035aa34e-kube-api-access-czmdn\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204543 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-system-cni-dir\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204563 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-openvswitch\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204579 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-etc-openvswitch\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204594 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-multus-socket-dir-parent\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204608 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-var-lib-openvswitch\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204622 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-run-ovn-kubernetes\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204636 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-run-netns\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.204650 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-etc-kubernetes\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.210408 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.221360 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.234107 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.247328 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.252822 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-68rwq" event={"ID":"3dfed96b-3244-4631-b080-cf112ad196dc","Type":"ContainerStarted","Data":"7377c9d3f61c1e9a8124b9f636a8dd84c98784e1e417d9bca65c10418f7e06d4"} Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.268095 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.295169 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.305885 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-multus-conf-dir\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.305923 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-var-lib-kubelet\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.305937 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-hostroot\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.305954 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-run-multus-certs\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.305970 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/39930cf4-5f3c-42cc-9384-d366bb1d6a78-multus-daemon-config\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.305986 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306004 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ea0cad1-8e8b-458c-a943-7191b36e35a3-mcd-auth-proxy-config\") pod \"machine-config-daemon-7mq9p\" (UID: \"4ea0cad1-8e8b-458c-a943-7191b36e35a3\") " pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306022 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-slash\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306021 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-multus-conf-dir\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306074 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-var-lib-cni-bin\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306038 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-var-lib-cni-bin\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306078 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-hostroot\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306113 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306149 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-slash\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306158 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306168 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-kubelet\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306206 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-run-multus-certs\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306280 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-kubelet\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306289 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-var-lib-kubelet\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306340 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-log-socket\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306389 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cknt9\" (UniqueName: \"kubernetes.io/projected/4ea0cad1-8e8b-458c-a943-7191b36e35a3-kube-api-access-cknt9\") pod \"machine-config-daemon-7mq9p\" (UID: \"4ea0cad1-8e8b-458c-a943-7191b36e35a3\") " pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306420 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-log-socket\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306437 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-multus-cni-dir\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306460 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-os-release\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306501 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306525 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65952\" (UniqueName: \"kubernetes.io/projected/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-kube-api-access-65952\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306550 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ea0cad1-8e8b-458c-a943-7191b36e35a3-proxy-tls\") pod \"machine-config-daemon-7mq9p\" (UID: \"4ea0cad1-8e8b-458c-a943-7191b36e35a3\") " pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306567 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-multus-cni-dir\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306572 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-cnibin\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306613 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4ea0cad1-8e8b-458c-a943-7191b36e35a3-rootfs\") pod \"machine-config-daemon-7mq9p\" (UID: \"4ea0cad1-8e8b-458c-a943-7191b36e35a3\") " pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306617 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-cnibin\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306632 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-node-log\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306650 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-cnibin\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306664 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-run-k8s-cni-cncf-io\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306681 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njxg4\" (UniqueName: \"kubernetes.io/projected/39930cf4-5f3c-42cc-9384-d366bb1d6a78-kube-api-access-njxg4\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306698 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovnkube-script-lib\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306709 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/39930cf4-5f3c-42cc-9384-d366bb1d6a78-multus-daemon-config\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306719 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-os-release\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306753 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-cnibin\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306760 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-env-overrides\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306778 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovn-node-metrics-cert\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306782 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/4ea0cad1-8e8b-458c-a943-7191b36e35a3-rootfs\") pod \"machine-config-daemon-7mq9p\" (UID: \"4ea0cad1-8e8b-458c-a943-7191b36e35a3\") " pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306793 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-systemd-units\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306805 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-node-log\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306816 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-cni-netd\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306838 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-system-cni-dir\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306846 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ea0cad1-8e8b-458c-a943-7191b36e35a3-mcd-auth-proxy-config\") pod \"machine-config-daemon-7mq9p\" (UID: \"4ea0cad1-8e8b-458c-a943-7191b36e35a3\") " pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306878 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czmdn\" (UniqueName: \"kubernetes.io/projected/e8ed6bbc-965a-412c-a3ee-4763035aa34e-kube-api-access-czmdn\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306921 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-etc-openvswitch\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306947 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-openvswitch\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306968 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-os-release\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.306972 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-var-lib-openvswitch\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307176 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307345 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-var-lib-openvswitch\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307351 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-systemd-units\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307370 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-cni-netd\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307353 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-multus-socket-dir-parent\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307399 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-run-k8s-cni-cncf-io\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307405 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-multus-socket-dir-parent\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307411 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-run-ovn-kubernetes\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307416 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-etc-openvswitch\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307434 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-openvswitch\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307468 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-run-ovn-kubernetes\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307470 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-run-netns\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307487 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-run-netns\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307488 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-os-release\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307519 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-etc-kubernetes\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307560 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-ovn\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307581 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-cni-bin\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307601 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-system-cni-dir\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307618 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-env-overrides\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307626 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-cni-bin\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307626 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-etc-kubernetes\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307621 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/39930cf4-5f3c-42cc-9384-d366bb1d6a78-cni-binary-copy\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307652 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-ovn\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307730 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovnkube-script-lib\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307762 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-run-netns\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307793 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-systemd\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307809 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-var-lib-cni-multus\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307839 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-cni-binary-copy\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307860 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/39930cf4-5f3c-42cc-9384-d366bb1d6a78-host-var-lib-cni-multus\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307875 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-run-netns\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.307889 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-systemd\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.308328 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-cni-binary-copy\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.308360 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovnkube-config\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.308392 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-system-cni-dir\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.308424 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/39930cf4-5f3c-42cc-9384-d366bb1d6a78-cni-binary-copy\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.308510 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-system-cni-dir\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.308805 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovnkube-config\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.309798 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.310843 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovn-node-metrics-cert\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.310855 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ea0cad1-8e8b-458c-a943-7191b36e35a3-proxy-tls\") pod \"machine-config-daemon-7mq9p\" (UID: \"4ea0cad1-8e8b-458c-a943-7191b36e35a3\") " pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.312581 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.326488 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cknt9\" (UniqueName: \"kubernetes.io/projected/4ea0cad1-8e8b-458c-a943-7191b36e35a3-kube-api-access-cknt9\") pod \"machine-config-daemon-7mq9p\" (UID: \"4ea0cad1-8e8b-458c-a943-7191b36e35a3\") " pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.327861 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65952\" (UniqueName: \"kubernetes.io/projected/92d21faf-b0b5-469f-8c6f-af3b0ee6240f-kube-api-access-65952\") pod \"multus-additional-cni-plugins-44ndp\" (UID: \"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\") " pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.330598 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.330855 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njxg4\" (UniqueName: \"kubernetes.io/projected/39930cf4-5f3c-42cc-9384-d366bb1d6a78-kube-api-access-njxg4\") pod \"multus-5cx6g\" (UID: \"39930cf4-5f3c-42cc-9384-d366bb1d6a78\") " pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.331089 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czmdn\" (UniqueName: \"kubernetes.io/projected/e8ed6bbc-965a-412c-a3ee-4763035aa34e-kube-api-access-czmdn\") pod \"ovnkube-node-rsgmm\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.344462 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.355951 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.366398 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.369488 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.369782 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.378439 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.383118 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.402889 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.418067 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.424179 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.436586 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-44ndp" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.438839 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.443192 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.457184 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5cx6g" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.457839 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.470877 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.482011 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.495196 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.498352 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.509799 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.515690 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.554599 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: W1127 10:21:58.554712 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92d21faf_b0b5_469f_8c6f_af3b0ee6240f.slice/crio-8e2db111c60135709f438372c6c85c37b216416aa038741e664821defa7c23c9 WatchSource:0}: Error finding container 8e2db111c60135709f438372c6c85c37b216416aa038741e664821defa7c23c9: Status 404 returned error can't find the container with id 8e2db111c60135709f438372c6c85c37b216416aa038741e664821defa7c23c9 Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.576196 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.593692 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.632837 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.652034 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.691161 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.725879 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.779637 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.801828 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.820841 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.838461 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.852764 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.871454 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.886776 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.906996 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.933505 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.950519 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.970374 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.983098 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:58 crc kubenswrapper[4806]: I1127 10:21:58.998177 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.012297 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.026667 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.038852 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.053686 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.085209 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.263700 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b"} Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.265200 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-68rwq" event={"ID":"3dfed96b-3244-4631-b080-cf112ad196dc","Type":"ContainerStarted","Data":"bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311"} Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.268481 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5cx6g" event={"ID":"39930cf4-5f3c-42cc-9384-d366bb1d6a78","Type":"ContainerStarted","Data":"6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2"} Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.268628 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5cx6g" event={"ID":"39930cf4-5f3c-42cc-9384-d366bb1d6a78","Type":"ContainerStarted","Data":"f6fef02355a25482db5fee2f750d78dffdf75cd450f7564df45723aed37a0f74"} Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.270876 4806 generic.go:334] "Generic (PLEG): container finished" podID="92d21faf-b0b5-469f-8c6f-af3b0ee6240f" containerID="137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1" exitCode=0 Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.270981 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" event={"ID":"92d21faf-b0b5-469f-8c6f-af3b0ee6240f","Type":"ContainerDied","Data":"137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1"} Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.271018 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" event={"ID":"92d21faf-b0b5-469f-8c6f-af3b0ee6240f","Type":"ContainerStarted","Data":"8e2db111c60135709f438372c6c85c37b216416aa038741e664821defa7c23c9"} Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.272554 4806 generic.go:334] "Generic (PLEG): container finished" podID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerID="2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961" exitCode=0 Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.272610 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961"} Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.272633 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerStarted","Data":"89943e205c0bc21f519cd22041759a0a025bfee3dfdc5e836e5939bf1b2b2232"} Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.275369 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5"} Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.275419 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218"} Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.275432 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"c7b7c39843c81237b077c64e08bb79bc5f80b6448006f11ce252070a8a7d13b4"} Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.280323 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.291287 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.314040 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.333999 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.359553 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.373037 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.386478 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.401834 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.418350 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.439542 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.462197 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.475926 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.489801 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.504264 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.517816 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.533442 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.558790 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.574253 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.587519 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.603202 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.624873 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:21:59 crc kubenswrapper[4806]: E1127 10:21:59.625086 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:21:59 crc kubenswrapper[4806]: E1127 10:21:59.625167 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:03.625150338 +0000 UTC m=+28.211741102 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.631613 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.655649 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.669006 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.684145 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.706959 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.725736 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:21:59 crc kubenswrapper[4806]: E1127 10:21:59.725829 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:22:03.725815059 +0000 UTC m=+28.312405823 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:21:59 crc kubenswrapper[4806]: E1127 10:21:59.726045 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:21:59 crc kubenswrapper[4806]: E1127 10:21:59.726063 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:21:59 crc kubenswrapper[4806]: E1127 10:21:59.726072 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.726452 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.726572 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.726613 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.726632 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:21:59 crc kubenswrapper[4806]: E1127 10:21:59.726660 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:03.726652349 +0000 UTC m=+28.313243113 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:59 crc kubenswrapper[4806]: E1127 10:21:59.726697 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:21:59 crc kubenswrapper[4806]: E1127 10:21:59.726777 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:21:59 crc kubenswrapper[4806]: E1127 10:21:59.726785 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:59 crc kubenswrapper[4806]: E1127 10:21:59.726806 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:03.726800322 +0000 UTC m=+28.313391076 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:21:59 crc kubenswrapper[4806]: E1127 10:21:59.726723 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:21:59 crc kubenswrapper[4806]: E1127 10:21:59.726830 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:03.726826083 +0000 UTC m=+28.313416847 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.743860 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:21:59 crc kubenswrapper[4806]: I1127 10:21:59.762913 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:21:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.116408 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:00 crc kubenswrapper[4806]: E1127 10:22:00.116863 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.116648 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.116608 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:00 crc kubenswrapper[4806]: E1127 10:22:00.117203 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:00 crc kubenswrapper[4806]: E1127 10:22:00.117006 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.282215 4806 generic.go:334] "Generic (PLEG): container finished" podID="92d21faf-b0b5-469f-8c6f-af3b0ee6240f" containerID="24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264" exitCode=0 Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.282285 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" event={"ID":"92d21faf-b0b5-469f-8c6f-af3b0ee6240f","Type":"ContainerDied","Data":"24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264"} Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.286878 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerStarted","Data":"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc"} Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.286951 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerStarted","Data":"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20"} Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.286973 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerStarted","Data":"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73"} Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.286992 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerStarted","Data":"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7"} Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.287009 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerStarted","Data":"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b"} Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.342267 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.380657 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.411733 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-dzlc2"] Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.412148 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dzlc2" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.414525 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.414737 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.415595 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.415973 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.426314 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.459026 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.483945 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.496288 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.509089 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.523970 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.534171 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfw9s\" (UniqueName: \"kubernetes.io/projected/9f8291c9-c70f-4d08-bb29-fbc523d38fa3-kube-api-access-hfw9s\") pod \"node-ca-dzlc2\" (UID: \"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\") " pod="openshift-image-registry/node-ca-dzlc2" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.534206 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f8291c9-c70f-4d08-bb29-fbc523d38fa3-host\") pod \"node-ca-dzlc2\" (UID: \"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\") " pod="openshift-image-registry/node-ca-dzlc2" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.534250 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9f8291c9-c70f-4d08-bb29-fbc523d38fa3-serviceca\") pod \"node-ca-dzlc2\" (UID: \"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\") " pod="openshift-image-registry/node-ca-dzlc2" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.538887 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.557329 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.568422 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.582058 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.592253 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.607881 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.626822 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.635418 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9f8291c9-c70f-4d08-bb29-fbc523d38fa3-serviceca\") pod \"node-ca-dzlc2\" (UID: \"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\") " pod="openshift-image-registry/node-ca-dzlc2" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.635488 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfw9s\" (UniqueName: \"kubernetes.io/projected/9f8291c9-c70f-4d08-bb29-fbc523d38fa3-kube-api-access-hfw9s\") pod \"node-ca-dzlc2\" (UID: \"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\") " pod="openshift-image-registry/node-ca-dzlc2" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.635508 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f8291c9-c70f-4d08-bb29-fbc523d38fa3-host\") pod \"node-ca-dzlc2\" (UID: \"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\") " pod="openshift-image-registry/node-ca-dzlc2" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.635752 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f8291c9-c70f-4d08-bb29-fbc523d38fa3-host\") pod \"node-ca-dzlc2\" (UID: \"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\") " pod="openshift-image-registry/node-ca-dzlc2" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.637109 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9f8291c9-c70f-4d08-bb29-fbc523d38fa3-serviceca\") pod \"node-ca-dzlc2\" (UID: \"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\") " pod="openshift-image-registry/node-ca-dzlc2" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.638437 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.651628 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.653190 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfw9s\" (UniqueName: \"kubernetes.io/projected/9f8291c9-c70f-4d08-bb29-fbc523d38fa3-kube-api-access-hfw9s\") pod \"node-ca-dzlc2\" (UID: \"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\") " pod="openshift-image-registry/node-ca-dzlc2" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.666014 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.676839 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.689788 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.703805 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.740612 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dzlc2" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.740603 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: W1127 10:22:00.752944 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f8291c9_c70f_4d08_bb29_fbc523d38fa3.slice/crio-7df61503279a6194f187f322973a0088603a7bd9bba12ae6c135040ea37293dc WatchSource:0}: Error finding container 7df61503279a6194f187f322973a0088603a7bd9bba12ae6c135040ea37293dc: Status 404 returned error can't find the container with id 7df61503279a6194f187f322973a0088603a7bd9bba12ae6c135040ea37293dc Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.777404 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.815424 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.859699 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.895544 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.935916 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:00 crc kubenswrapper[4806]: I1127 10:22:00.975273 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:00Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.022046 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.292031 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerStarted","Data":"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744"} Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.293691 4806 generic.go:334] "Generic (PLEG): container finished" podID="92d21faf-b0b5-469f-8c6f-af3b0ee6240f" containerID="fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c" exitCode=0 Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.293749 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" event={"ID":"92d21faf-b0b5-469f-8c6f-af3b0ee6240f","Type":"ContainerDied","Data":"fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c"} Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.294965 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dzlc2" event={"ID":"9f8291c9-c70f-4d08-bb29-fbc523d38fa3","Type":"ContainerStarted","Data":"c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20"} Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.294995 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dzlc2" event={"ID":"9f8291c9-c70f-4d08-bb29-fbc523d38fa3","Type":"ContainerStarted","Data":"7df61503279a6194f187f322973a0088603a7bd9bba12ae6c135040ea37293dc"} Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.307258 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.321705 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.339340 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.353404 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.372547 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.387266 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.401279 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.418000 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.428670 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.437905 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.457376 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.494093 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.539020 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.584772 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.617005 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.654718 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.703166 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.737744 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.775553 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.819992 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.856178 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.873284 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.876133 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.876178 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.876189 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.876388 4806 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.898972 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.949836 4806 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.950224 4806 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.952031 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.952091 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.952118 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.952151 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.952301 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:01Z","lastTransitionTime":"2025-11-27T10:22:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:01 crc kubenswrapper[4806]: E1127 10:22:01.974692 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.976110 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.978842 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.978891 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.978902 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.978920 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.978930 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:01Z","lastTransitionTime":"2025-11-27T10:22:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:01 crc kubenswrapper[4806]: E1127 10:22:01.992948 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:01Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.996921 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.996978 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.996990 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.997013 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:01 crc kubenswrapper[4806]: I1127 10:22:01.997026 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:01Z","lastTransitionTime":"2025-11-27T10:22:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:02 crc kubenswrapper[4806]: E1127 10:22:02.010043 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.014011 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.014045 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.014054 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.014069 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.014078 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:02Z","lastTransitionTime":"2025-11-27T10:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.028904 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: E1127 10:22:02.031445 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.036554 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.036623 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.036640 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.036665 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.036681 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:02Z","lastTransitionTime":"2025-11-27T10:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:02 crc kubenswrapper[4806]: E1127 10:22:02.052640 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: E1127 10:22:02.052828 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.055085 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.055122 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.055134 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.055159 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.055172 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:02Z","lastTransitionTime":"2025-11-27T10:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.065083 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.108553 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.115838 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:02 crc kubenswrapper[4806]: E1127 10:22:02.116024 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.116104 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.116162 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:02 crc kubenswrapper[4806]: E1127 10:22:02.116254 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:02 crc kubenswrapper[4806]: E1127 10:22:02.116316 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.158047 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.158103 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.158114 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.158136 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.158148 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:02Z","lastTransitionTime":"2025-11-27T10:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.178601 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.195263 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.219082 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.258202 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.261256 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.261315 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.261328 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.261350 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.261364 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:02Z","lastTransitionTime":"2025-11-27T10:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.300655 4806 generic.go:334] "Generic (PLEG): container finished" podID="92d21faf-b0b5-469f-8c6f-af3b0ee6240f" containerID="ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e" exitCode=0 Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.300696 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" event={"ID":"92d21faf-b0b5-469f-8c6f-af3b0ee6240f","Type":"ContainerDied","Data":"ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e"} Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.322879 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.342838 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.366195 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.366247 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.366258 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.366272 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.366281 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:02Z","lastTransitionTime":"2025-11-27T10:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.378047 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.415511 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.459015 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.469956 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.470009 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.470026 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.470055 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.470071 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:02Z","lastTransitionTime":"2025-11-27T10:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.499772 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.540968 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.572591 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.572633 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.572645 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.572663 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.572675 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:02Z","lastTransitionTime":"2025-11-27T10:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.576114 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.616970 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.663273 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.674402 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.674439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.674465 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.674479 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.674490 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:02Z","lastTransitionTime":"2025-11-27T10:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.697497 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.735957 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.776370 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.776420 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.776432 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.776451 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.776464 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:02Z","lastTransitionTime":"2025-11-27T10:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.779431 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.822396 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.855298 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:02Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.879080 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.879111 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.879121 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.879136 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.879148 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:02Z","lastTransitionTime":"2025-11-27T10:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.981156 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.981195 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.981206 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.981223 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:02 crc kubenswrapper[4806]: I1127 10:22:02.981252 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:02Z","lastTransitionTime":"2025-11-27T10:22:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.083130 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.083172 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.083182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.083198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.083209 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:03Z","lastTransitionTime":"2025-11-27T10:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.186182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.186252 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.186271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.186293 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.186307 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:03Z","lastTransitionTime":"2025-11-27T10:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.288977 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.289010 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.289021 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.289033 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.289042 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:03Z","lastTransitionTime":"2025-11-27T10:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.310102 4806 generic.go:334] "Generic (PLEG): container finished" podID="92d21faf-b0b5-469f-8c6f-af3b0ee6240f" containerID="011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7" exitCode=0 Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.310157 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" event={"ID":"92d21faf-b0b5-469f-8c6f-af3b0ee6240f","Type":"ContainerDied","Data":"011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7"} Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.317557 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerStarted","Data":"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef"} Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.326488 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.343709 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.355419 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.367927 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.386484 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.390832 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.390938 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.390974 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.391020 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.391035 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:03Z","lastTransitionTime":"2025-11-27T10:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.402790 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.415995 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.433969 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.445343 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.459314 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.473401 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.485008 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.494474 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.494515 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.494528 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.494575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.494591 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:03Z","lastTransitionTime":"2025-11-27T10:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.504990 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.517404 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.532891 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.597427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.597457 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.597467 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.597479 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.597487 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:03Z","lastTransitionTime":"2025-11-27T10:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.677833 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:03 crc kubenswrapper[4806]: E1127 10:22:03.678163 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:22:03 crc kubenswrapper[4806]: E1127 10:22:03.678366 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:11.678305358 +0000 UTC m=+36.264896162 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.700606 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.700637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.700645 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.700657 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.700666 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:03Z","lastTransitionTime":"2025-11-27T10:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.702889 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.720886 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.740334 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.754928 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.775330 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.778434 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.778533 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.778566 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.778592 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:03 crc kubenswrapper[4806]: E1127 10:22:03.778688 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:22:11.778653601 +0000 UTC m=+36.365244395 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:22:03 crc kubenswrapper[4806]: E1127 10:22:03.778724 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:22:03 crc kubenswrapper[4806]: E1127 10:22:03.778726 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:22:03 crc kubenswrapper[4806]: E1127 10:22:03.778734 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:22:03 crc kubenswrapper[4806]: E1127 10:22:03.778762 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:22:03 crc kubenswrapper[4806]: E1127 10:22:03.778827 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:03 crc kubenswrapper[4806]: E1127 10:22:03.778855 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:11.778837046 +0000 UTC m=+36.365427850 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:22:03 crc kubenswrapper[4806]: E1127 10:22:03.778740 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:22:03 crc kubenswrapper[4806]: E1127 10:22:03.778880 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:03 crc kubenswrapper[4806]: E1127 10:22:03.778883 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:11.778870976 +0000 UTC m=+36.365461770 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:03 crc kubenswrapper[4806]: E1127 10:22:03.778937 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:11.778919317 +0000 UTC m=+36.365510111 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.793395 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.803586 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.803632 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.803644 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.803660 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.803671 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:03Z","lastTransitionTime":"2025-11-27T10:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.814067 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.837069 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.851424 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.868166 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.883482 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.906940 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.906977 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.906988 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.907004 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.907015 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:03Z","lastTransitionTime":"2025-11-27T10:22:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.908940 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.943470 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:03 crc kubenswrapper[4806]: I1127 10:22:03.984747 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.009380 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.009443 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.009470 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.009505 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.009531 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:04Z","lastTransitionTime":"2025-11-27T10:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.023753 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.066886 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.114284 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.114349 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.114367 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.114396 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.114416 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:04Z","lastTransitionTime":"2025-11-27T10:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.116206 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.116200 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.116332 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:04 crc kubenswrapper[4806]: E1127 10:22:04.116500 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:04 crc kubenswrapper[4806]: E1127 10:22:04.116680 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:04 crc kubenswrapper[4806]: E1127 10:22:04.117195 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.254721 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.254780 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.254792 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.254810 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.254837 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:04Z","lastTransitionTime":"2025-11-27T10:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.323632 4806 generic.go:334] "Generic (PLEG): container finished" podID="92d21faf-b0b5-469f-8c6f-af3b0ee6240f" containerID="5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d" exitCode=0 Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.323680 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" event={"ID":"92d21faf-b0b5-469f-8c6f-af3b0ee6240f","Type":"ContainerDied","Data":"5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d"} Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.339519 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.349668 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.357207 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.357243 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.357256 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.357269 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.357278 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:04Z","lastTransitionTime":"2025-11-27T10:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.364271 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.425370 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.435509 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.445760 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.459752 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.459781 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.459789 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.459802 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.459811 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:04Z","lastTransitionTime":"2025-11-27T10:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.463627 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.477219 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.495028 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.517362 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.538265 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.559512 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.564295 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.564334 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.564345 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.564359 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.564367 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:04Z","lastTransitionTime":"2025-11-27T10:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.589488 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.619802 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.675585 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:04Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.677549 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.677579 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.677590 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.677605 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.677615 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:04Z","lastTransitionTime":"2025-11-27T10:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.780020 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.780092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.780108 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.780133 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.780147 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:04Z","lastTransitionTime":"2025-11-27T10:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.882965 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.883003 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.883014 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.883031 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.883049 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:04Z","lastTransitionTime":"2025-11-27T10:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.986351 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.986419 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.986437 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.986459 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:04 crc kubenswrapper[4806]: I1127 10:22:04.986471 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:04Z","lastTransitionTime":"2025-11-27T10:22:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.089481 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.089543 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.089563 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.089588 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.089602 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:05Z","lastTransitionTime":"2025-11-27T10:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.192523 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.192577 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.192590 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.192609 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.192622 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:05Z","lastTransitionTime":"2025-11-27T10:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.295825 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.295863 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.295873 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.295889 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.295897 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:05Z","lastTransitionTime":"2025-11-27T10:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.361824 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" event={"ID":"92d21faf-b0b5-469f-8c6f-af3b0ee6240f","Type":"ContainerStarted","Data":"b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac"} Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.366394 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerStarted","Data":"68af93dc72eeab13d19450a150ac003f3138658dca28e9699c41e124690e063a"} Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.366779 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.376027 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.386844 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.398487 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.398523 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.398531 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.398548 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.398558 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:05Z","lastTransitionTime":"2025-11-27T10:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.401576 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.405536 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.415362 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.436179 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.448929 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.466125 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.479956 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.500826 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.500873 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.500889 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.500908 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.500920 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:05Z","lastTransitionTime":"2025-11-27T10:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.504612 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.519963 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.543041 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.555921 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.573602 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.587217 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.599616 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.603708 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.603761 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.603776 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.603792 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.603804 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:05Z","lastTransitionTime":"2025-11-27T10:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.614001 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.628309 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.645326 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.656578 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.675269 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68af93dc72eeab13d19450a150ac003f3138658dca28e9699c41e124690e063a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.690945 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.706267 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.706324 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.706337 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.706363 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.706375 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:05Z","lastTransitionTime":"2025-11-27T10:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.708057 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.722818 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.741425 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.753411 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.775328 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.787630 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.800366 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.813341 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.813408 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.813422 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.813448 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.813466 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:05Z","lastTransitionTime":"2025-11-27T10:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.824199 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.860215 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:05Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.916365 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.916435 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.916460 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.916490 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:05 crc kubenswrapper[4806]: I1127 10:22:05.916510 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:05Z","lastTransitionTime":"2025-11-27T10:22:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.020248 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.020289 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.020314 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.020340 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.020355 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:06Z","lastTransitionTime":"2025-11-27T10:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.116092 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.116165 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.116212 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:06 crc kubenswrapper[4806]: E1127 10:22:06.116318 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:06 crc kubenswrapper[4806]: E1127 10:22:06.116430 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:06 crc kubenswrapper[4806]: E1127 10:22:06.116628 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.123198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.123279 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.123297 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.123324 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.123339 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:06Z","lastTransitionTime":"2025-11-27T10:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.152482 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68af93dc72eeab13d19450a150ac003f3138658dca28e9699c41e124690e063a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.170176 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.189784 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.209857 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.225061 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.225109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.225126 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.225158 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.225175 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:06Z","lastTransitionTime":"2025-11-27T10:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.235689 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.254921 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.270421 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.293168 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.309275 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.327198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.327273 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.327286 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.327305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.327324 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:06Z","lastTransitionTime":"2025-11-27T10:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.328222 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.347653 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.365092 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.369779 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.370193 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.386904 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.436990 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.437446 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.437558 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.437651 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.437727 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:06Z","lastTransitionTime":"2025-11-27T10:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.442155 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.446845 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.465892 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.497338 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.539929 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.539989 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.540005 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.540026 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.540045 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:06Z","lastTransitionTime":"2025-11-27T10:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.541041 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.581522 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.623072 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.646180 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.646617 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.646689 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.646753 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.646807 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:06Z","lastTransitionTime":"2025-11-27T10:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.666385 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.708329 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.749282 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.751516 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.751675 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.751840 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.753960 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.754043 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:06Z","lastTransitionTime":"2025-11-27T10:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.800973 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68af93dc72eeab13d19450a150ac003f3138658dca28e9699c41e124690e063a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.823448 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.856625 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.856680 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.856696 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.856719 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.856736 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:06Z","lastTransitionTime":"2025-11-27T10:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.862206 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.898592 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.940173 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.958969 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.959410 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.959498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.959580 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.959650 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:06Z","lastTransitionTime":"2025-11-27T10:22:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:06 crc kubenswrapper[4806]: I1127 10:22:06.980013 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.019587 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:07Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.060430 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:07Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.062271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.062322 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.062334 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.062394 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.062409 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:07Z","lastTransitionTime":"2025-11-27T10:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.170611 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.170651 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.170663 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.170680 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.170692 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:07Z","lastTransitionTime":"2025-11-27T10:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.275522 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.275576 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.275593 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.275612 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.275623 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:07Z","lastTransitionTime":"2025-11-27T10:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.372148 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.377369 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.377417 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.377427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.377447 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.377461 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:07Z","lastTransitionTime":"2025-11-27T10:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.480023 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.480058 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.480073 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.480092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.480107 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:07Z","lastTransitionTime":"2025-11-27T10:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.583209 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.583319 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.583342 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.583362 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.583373 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:07Z","lastTransitionTime":"2025-11-27T10:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.685408 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.685442 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.685452 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.685470 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.685480 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:07Z","lastTransitionTime":"2025-11-27T10:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.790653 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.790725 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.790738 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.790757 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.790770 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:07Z","lastTransitionTime":"2025-11-27T10:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.893369 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.893411 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.893420 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.893436 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.893447 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:07Z","lastTransitionTime":"2025-11-27T10:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.997566 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.997633 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.997653 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.997688 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:07 crc kubenswrapper[4806]: I1127 10:22:07.997724 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:07Z","lastTransitionTime":"2025-11-27T10:22:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.100920 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.100963 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.100979 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.100996 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.101007 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:08Z","lastTransitionTime":"2025-11-27T10:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.116675 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:08 crc kubenswrapper[4806]: E1127 10:22:08.116918 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.117600 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:08 crc kubenswrapper[4806]: E1127 10:22:08.117717 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.117858 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:08 crc kubenswrapper[4806]: E1127 10:22:08.118037 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.203675 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.203712 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.203719 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.203733 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.203741 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:08Z","lastTransitionTime":"2025-11-27T10:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.307137 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.307175 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.307184 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.307198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.307208 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:08Z","lastTransitionTime":"2025-11-27T10:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.380463 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/0.log" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.384425 4806 generic.go:334] "Generic (PLEG): container finished" podID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerID="68af93dc72eeab13d19450a150ac003f3138658dca28e9699c41e124690e063a" exitCode=1 Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.384462 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"68af93dc72eeab13d19450a150ac003f3138658dca28e9699c41e124690e063a"} Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.385080 4806 scope.go:117] "RemoveContainer" containerID="68af93dc72eeab13d19450a150ac003f3138658dca28e9699c41e124690e063a" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.405961 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.410456 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.410925 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.411431 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.411544 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.411638 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:08Z","lastTransitionTime":"2025-11-27T10:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.422834 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.458695 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68af93dc72eeab13d19450a150ac003f3138658dca28e9699c41e124690e063a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68af93dc72eeab13d19450a150ac003f3138658dca28e9699c41e124690e063a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:08Z\\\",\\\"message\\\":\\\"07.857667 6005 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1127 10:22:07.857690 6005 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1127 10:22:07.857702 6005 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 10:22:07.857733 6005 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1127 10:22:07.857778 6005 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 10:22:07.857848 6005 factory.go:656] Stopping watch factory\\\\nI1127 10:22:07.857883 6005 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 10:22:07.857898 6005 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 10:22:07.857907 6005 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1127 10:22:07.857916 6005 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 10:22:07.857925 6005 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 10:22:07.857934 6005 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1127 10:22:07.858092 6005 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 10:22:07.858259 6005 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.475656 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.491707 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.502017 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.514943 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.515642 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.515682 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.515695 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.515713 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.515725 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:08Z","lastTransitionTime":"2025-11-27T10:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.524298 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.541352 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.552781 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.567051 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.579484 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.591314 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.603825 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.616014 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:08Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.617592 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.617626 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.617639 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.617661 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.617675 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:08Z","lastTransitionTime":"2025-11-27T10:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.720074 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.720127 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.720144 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.720172 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.720191 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:08Z","lastTransitionTime":"2025-11-27T10:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.823108 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.823162 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.823176 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.823201 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.823217 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:08Z","lastTransitionTime":"2025-11-27T10:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.925047 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.925084 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.925093 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.925108 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:08 crc kubenswrapper[4806]: I1127 10:22:08.925117 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:08Z","lastTransitionTime":"2025-11-27T10:22:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.027517 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.027573 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.027591 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.027616 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.027635 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:09Z","lastTransitionTime":"2025-11-27T10:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.130669 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.130707 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.130716 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.130731 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.130740 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:09Z","lastTransitionTime":"2025-11-27T10:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.233522 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.233561 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.233582 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.233600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.233612 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:09Z","lastTransitionTime":"2025-11-27T10:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.337894 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.337968 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.337992 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.338021 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.338044 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:09Z","lastTransitionTime":"2025-11-27T10:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.389575 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/1.log" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.390211 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/0.log" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.392932 4806 generic.go:334] "Generic (PLEG): container finished" podID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerID="29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a" exitCode=1 Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.392967 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a"} Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.392997 4806 scope.go:117] "RemoveContainer" containerID="68af93dc72eeab13d19450a150ac003f3138658dca28e9699c41e124690e063a" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.394938 4806 scope.go:117] "RemoveContainer" containerID="29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a" Nov 27 10:22:09 crc kubenswrapper[4806]: E1127 10:22:09.395202 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.410039 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.422396 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.439896 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.440751 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.440774 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.440781 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.440793 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.440802 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:09Z","lastTransitionTime":"2025-11-27T10:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.450482 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.462371 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.475282 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.493496 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.507437 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.519523 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.536959 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68af93dc72eeab13d19450a150ac003f3138658dca28e9699c41e124690e063a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:08Z\\\",\\\"message\\\":\\\"07.857667 6005 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1127 10:22:07.857690 6005 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1127 10:22:07.857702 6005 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 10:22:07.857733 6005 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1127 10:22:07.857778 6005 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 10:22:07.857848 6005 factory.go:656] Stopping watch factory\\\\nI1127 10:22:07.857883 6005 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 10:22:07.857898 6005 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 10:22:07.857907 6005 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1127 10:22:07.857916 6005 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 10:22:07.857925 6005 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 10:22:07.857934 6005 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1127 10:22:07.858092 6005 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 10:22:07.858259 6005 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"w:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1127 10:22:09.261320 6125 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.543579 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.543645 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.543667 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.543696 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.543720 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:09Z","lastTransitionTime":"2025-11-27T10:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.550857 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.560164 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.570318 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.583530 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.596124 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.645872 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.646290 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.646443 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.646570 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.646721 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:09Z","lastTransitionTime":"2025-11-27T10:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.751102 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.751149 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.751166 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.751187 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.751206 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:09Z","lastTransitionTime":"2025-11-27T10:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.854698 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.854747 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.854770 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.854798 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.854821 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:09Z","lastTransitionTime":"2025-11-27T10:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.870111 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n"] Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.871097 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.875255 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.876596 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.897516 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.931947 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.947885 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.957256 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.957285 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.957298 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.957319 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.957335 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:09Z","lastTransitionTime":"2025-11-27T10:22:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.967053 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.973504 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e5d1ac6b-daa1-4c04-a3b3-51ce910ab441-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lqb5n\" (UID: \"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.973744 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2jnn\" (UniqueName: \"kubernetes.io/projected/e5d1ac6b-daa1-4c04-a3b3-51ce910ab441-kube-api-access-g2jnn\") pod \"ovnkube-control-plane-749d76644c-lqb5n\" (UID: \"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.973891 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e5d1ac6b-daa1-4c04-a3b3-51ce910ab441-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lqb5n\" (UID: \"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.974026 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e5d1ac6b-daa1-4c04-a3b3-51ce910ab441-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lqb5n\" (UID: \"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.981541 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:09 crc kubenswrapper[4806]: I1127 10:22:09.995619 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:09Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.007265 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.022219 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.031006 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.047816 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68af93dc72eeab13d19450a150ac003f3138658dca28e9699c41e124690e063a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:08Z\\\",\\\"message\\\":\\\"07.857667 6005 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1127 10:22:07.857690 6005 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1127 10:22:07.857702 6005 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1127 10:22:07.857733 6005 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1127 10:22:07.857778 6005 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1127 10:22:07.857848 6005 factory.go:656] Stopping watch factory\\\\nI1127 10:22:07.857883 6005 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1127 10:22:07.857898 6005 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1127 10:22:07.857907 6005 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1127 10:22:07.857916 6005 handler.go:208] Removed *v1.Node event handler 2\\\\nI1127 10:22:07.857925 6005 handler.go:208] Removed *v1.Node event handler 7\\\\nI1127 10:22:07.857934 6005 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1127 10:22:07.858092 6005 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1127 10:22:07.858259 6005 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"w:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1127 10:22:09.261320 6125 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.059050 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.059317 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.059429 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.059544 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.059714 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:10Z","lastTransitionTime":"2025-11-27T10:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.060693 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.075362 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e5d1ac6b-daa1-4c04-a3b3-51ce910ab441-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lqb5n\" (UID: \"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.075407 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2jnn\" (UniqueName: \"kubernetes.io/projected/e5d1ac6b-daa1-4c04-a3b3-51ce910ab441-kube-api-access-g2jnn\") pod \"ovnkube-control-plane-749d76644c-lqb5n\" (UID: \"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.075437 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e5d1ac6b-daa1-4c04-a3b3-51ce910ab441-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lqb5n\" (UID: \"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.075462 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e5d1ac6b-daa1-4c04-a3b3-51ce910ab441-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lqb5n\" (UID: \"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.076610 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e5d1ac6b-daa1-4c04-a3b3-51ce910ab441-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lqb5n\" (UID: \"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.077493 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e5d1ac6b-daa1-4c04-a3b3-51ce910ab441-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lqb5n\" (UID: \"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.085903 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e5d1ac6b-daa1-4c04-a3b3-51ce910ab441-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lqb5n\" (UID: \"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.104956 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2jnn\" (UniqueName: \"kubernetes.io/projected/e5d1ac6b-daa1-4c04-a3b3-51ce910ab441-kube-api-access-g2jnn\") pod \"ovnkube-control-plane-749d76644c-lqb5n\" (UID: \"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.106689 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.115867 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:10 crc kubenswrapper[4806]: E1127 10:22:10.115997 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.115890 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:10 crc kubenswrapper[4806]: E1127 10:22:10.116072 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.115873 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:10 crc kubenswrapper[4806]: E1127 10:22:10.116122 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.139404 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.161043 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.161998 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.162028 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.162038 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.162053 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.162063 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:10Z","lastTransitionTime":"2025-11-27T10:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.172493 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.186435 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.202547 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.264769 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.264793 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.264801 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.264813 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.264821 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:10Z","lastTransitionTime":"2025-11-27T10:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.367424 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.367450 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.367460 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.367473 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.367482 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:10Z","lastTransitionTime":"2025-11-27T10:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.398303 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/1.log" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.401006 4806 scope.go:117] "RemoveContainer" containerID="29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a" Nov 27 10:22:10 crc kubenswrapper[4806]: E1127 10:22:10.401167 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.401201 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" event={"ID":"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441","Type":"ContainerStarted","Data":"637f9255a4aff397c3d1e79cc02674b0988c0b3f520e47dfbd656ac29c302f94"} Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.413665 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.426362 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.439600 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.449054 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.467875 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"w:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1127 10:22:09.261320 6125 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.470084 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.470113 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.470121 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.470134 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.470143 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:10Z","lastTransitionTime":"2025-11-27T10:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.479495 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.498436 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.519728 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.533563 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.556648 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.569354 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.572918 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.573014 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.573027 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.573043 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.573072 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:10Z","lastTransitionTime":"2025-11-27T10:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.590801 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.603512 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.623283 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.638564 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.654520 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:10Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.678077 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.678454 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.678478 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.678495 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.678511 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:10Z","lastTransitionTime":"2025-11-27T10:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.781302 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.781339 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.781352 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.781366 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.781377 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:10Z","lastTransitionTime":"2025-11-27T10:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.884131 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.884174 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.884184 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.884198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.884257 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:10Z","lastTransitionTime":"2025-11-27T10:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.986725 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.986781 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.986793 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.986811 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:10 crc kubenswrapper[4806]: I1127 10:22:10.986821 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:10Z","lastTransitionTime":"2025-11-27T10:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.089637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.089688 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.089701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.089718 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.089730 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:11Z","lastTransitionTime":"2025-11-27T10:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.192905 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.192960 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.192971 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.192985 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.192994 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:11Z","lastTransitionTime":"2025-11-27T10:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.238278 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.296938 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.296996 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.297012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.297035 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.297052 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:11Z","lastTransitionTime":"2025-11-27T10:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.399709 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.399752 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.399766 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.399790 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.399804 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:11Z","lastTransitionTime":"2025-11-27T10:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.406459 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" event={"ID":"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441","Type":"ContainerStarted","Data":"81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34"} Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.406529 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" event={"ID":"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441","Type":"ContainerStarted","Data":"9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21"} Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.407016 4806 scope.go:117] "RemoveContainer" containerID="29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a" Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.407693 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.429366 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.447688 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.468298 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.482657 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.499179 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.502495 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.502542 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.502564 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.502607 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.502631 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:11Z","lastTransitionTime":"2025-11-27T10:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.512438 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.527460 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.543403 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.561441 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.582287 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.597507 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.605147 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.605220 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.605251 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.605264 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.605273 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:11Z","lastTransitionTime":"2025-11-27T10:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.611510 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.629196 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"w:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1127 10:22:09.261320 6125 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.641567 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.654683 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.665256 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.687821 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.688011 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.688096 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:27.688074781 +0000 UTC m=+52.274665535 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.707872 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.707926 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.707942 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.707963 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.707976 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:11Z","lastTransitionTime":"2025-11-27T10:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.769869 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-p2gpw"] Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.770510 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.770599 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.789072 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.789211 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.789264 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.789285 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.789352 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.789426 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.789438 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.789447 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.789455 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:22:27.789418181 +0000 UTC m=+52.376008955 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.789499 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:27.789483712 +0000 UTC m=+52.376074596 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.789513 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.789522 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.789531 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.789555 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:27.789543304 +0000 UTC m=+52.376134168 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.789578 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:27.789567695 +0000 UTC m=+52.376158569 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.796489 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.810099 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.810164 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.810177 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.810195 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.810208 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:11Z","lastTransitionTime":"2025-11-27T10:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.811870 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.823885 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.836250 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.847562 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.867083 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.881494 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.889889 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrmz9\" (UniqueName: \"kubernetes.io/projected/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-kube-api-access-lrmz9\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.889964 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.892194 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.912589 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.912618 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.912626 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.912642 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.912651 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:11Z","lastTransitionTime":"2025-11-27T10:22:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.915114 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.927167 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.944223 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.958316 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.975328 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.991010 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrmz9\" (UniqueName: \"kubernetes.io/projected/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-kube-api-access-lrmz9\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.991092 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:11 crc kubenswrapper[4806]: I1127 10:22:11.991056 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:11Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.991205 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:11 crc kubenswrapper[4806]: E1127 10:22:11.991291 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs podName:3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e nodeName:}" failed. No retries permitted until 2025-11-27 10:22:12.491272426 +0000 UTC m=+37.077863200 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs") pod "network-metrics-daemon-p2gpw" (UID: "3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.007979 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:12Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.015365 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.015432 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.015454 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.015482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.015503 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.020483 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrmz9\" (UniqueName: \"kubernetes.io/projected/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-kube-api-access-lrmz9\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.039638 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"w:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1127 10:22:09.261320 6125 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:12Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.052751 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:12Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.116036 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.116040 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:12 crc kubenswrapper[4806]: E1127 10:22:12.116178 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:12 crc kubenswrapper[4806]: E1127 10:22:12.116325 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.116568 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:12 crc kubenswrapper[4806]: E1127 10:22:12.116765 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.117767 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.117809 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.117826 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.117845 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.117860 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.220711 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.220745 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.220753 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.220766 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.220775 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.249325 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.249367 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.249379 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.249393 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.249404 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: E1127 10:22:12.262558 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:12Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.268619 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.268659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.268671 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.268687 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.268699 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: E1127 10:22:12.282680 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:12Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.286462 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.286519 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.286536 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.286557 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.286574 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: E1127 10:22:12.301631 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:12Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.306113 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.306345 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.306370 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.306388 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.306424 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: E1127 10:22:12.322572 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:12Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.327937 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.327988 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.328001 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.328020 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.328034 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: E1127 10:22:12.343964 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:12Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:12 crc kubenswrapper[4806]: E1127 10:22:12.344130 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.346131 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.346171 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.346184 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.346210 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.346223 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.449061 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.449118 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.449134 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.449158 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.449175 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.495902 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:12 crc kubenswrapper[4806]: E1127 10:22:12.497427 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:12 crc kubenswrapper[4806]: E1127 10:22:12.497599 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs podName:3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e nodeName:}" failed. No retries permitted until 2025-11-27 10:22:13.49757804 +0000 UTC m=+38.084168804 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs") pod "network-metrics-daemon-p2gpw" (UID: "3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.552282 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.552343 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.552357 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.552375 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.552387 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.656304 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.656683 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.656710 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.656741 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.656764 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.761187 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.761280 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.761309 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.761341 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.761364 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.866223 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.866328 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.866350 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.866378 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.866398 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.971498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.971566 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.971591 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.971615 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:12 crc kubenswrapper[4806]: I1127 10:22:12.971633 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:12Z","lastTransitionTime":"2025-11-27T10:22:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.074893 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.074955 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.074971 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.074994 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.075013 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:13Z","lastTransitionTime":"2025-11-27T10:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.115677 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:13 crc kubenswrapper[4806]: E1127 10:22:13.116401 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.178113 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.178263 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.178283 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.178306 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.178359 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:13Z","lastTransitionTime":"2025-11-27T10:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.280810 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.280862 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.280879 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.280902 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.280918 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:13Z","lastTransitionTime":"2025-11-27T10:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.383750 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.383832 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.383850 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.383873 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.383890 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:13Z","lastTransitionTime":"2025-11-27T10:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.486493 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.486556 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.486575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.486601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.486617 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:13Z","lastTransitionTime":"2025-11-27T10:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.506330 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:13 crc kubenswrapper[4806]: E1127 10:22:13.506462 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:13 crc kubenswrapper[4806]: E1127 10:22:13.506516 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs podName:3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e nodeName:}" failed. No retries permitted until 2025-11-27 10:22:15.506501769 +0000 UTC m=+40.093092533 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs") pod "network-metrics-daemon-p2gpw" (UID: "3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.588949 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.589005 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.589023 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.589045 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.589065 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:13Z","lastTransitionTime":"2025-11-27T10:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.691446 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.691509 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.691532 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.691564 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.691585 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:13Z","lastTransitionTime":"2025-11-27T10:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.795111 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.795176 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.795198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.795227 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.795285 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:13Z","lastTransitionTime":"2025-11-27T10:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.899049 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.899108 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.899125 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.899151 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:13 crc kubenswrapper[4806]: I1127 10:22:13.899168 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:13Z","lastTransitionTime":"2025-11-27T10:22:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.002587 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.002640 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.002655 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.002678 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.002694 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:14Z","lastTransitionTime":"2025-11-27T10:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.105586 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.105639 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.105657 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.105679 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.105695 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:14Z","lastTransitionTime":"2025-11-27T10:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.116384 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.116381 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.116491 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:14 crc kubenswrapper[4806]: E1127 10:22:14.116701 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:14 crc kubenswrapper[4806]: E1127 10:22:14.117049 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:14 crc kubenswrapper[4806]: E1127 10:22:14.117212 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.209012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.209080 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.209092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.209111 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.209145 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:14Z","lastTransitionTime":"2025-11-27T10:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.312627 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.312689 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.312708 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.312733 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.312753 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:14Z","lastTransitionTime":"2025-11-27T10:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.417562 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.417626 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.417649 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.417682 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.417706 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:14Z","lastTransitionTime":"2025-11-27T10:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.519999 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.520061 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.520086 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.520117 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.520139 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:14Z","lastTransitionTime":"2025-11-27T10:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.623175 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.623287 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.623310 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.623338 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.623360 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:14Z","lastTransitionTime":"2025-11-27T10:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.726405 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.726440 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.726448 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.726463 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.726472 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:14Z","lastTransitionTime":"2025-11-27T10:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.829319 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.829390 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.829407 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.829429 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.829446 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:14Z","lastTransitionTime":"2025-11-27T10:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.933660 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.933751 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.933778 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.934198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:14 crc kubenswrapper[4806]: I1127 10:22:14.934216 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:14Z","lastTransitionTime":"2025-11-27T10:22:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.037681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.037728 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.037740 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.037757 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.037772 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:15Z","lastTransitionTime":"2025-11-27T10:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.116538 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:15 crc kubenswrapper[4806]: E1127 10:22:15.116779 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.141995 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.142065 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.142089 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.142119 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.142141 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:15Z","lastTransitionTime":"2025-11-27T10:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.245789 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.245884 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.245906 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.245930 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.245949 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:15Z","lastTransitionTime":"2025-11-27T10:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.349119 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.349165 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.349182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.349201 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.349215 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:15Z","lastTransitionTime":"2025-11-27T10:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.452082 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.452159 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.452181 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.452212 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.452266 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:15Z","lastTransitionTime":"2025-11-27T10:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.530112 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:15 crc kubenswrapper[4806]: E1127 10:22:15.530369 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:15 crc kubenswrapper[4806]: E1127 10:22:15.530481 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs podName:3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e nodeName:}" failed. No retries permitted until 2025-11-27 10:22:19.530448154 +0000 UTC m=+44.117038958 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs") pod "network-metrics-daemon-p2gpw" (UID: "3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.555855 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.555895 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.555908 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.555927 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.555940 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:15Z","lastTransitionTime":"2025-11-27T10:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.659353 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.659438 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.659455 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.659482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.659499 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:15Z","lastTransitionTime":"2025-11-27T10:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.763297 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.763348 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.763360 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.763379 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.763390 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:15Z","lastTransitionTime":"2025-11-27T10:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.865803 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.865840 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.865854 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.865870 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.865882 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:15Z","lastTransitionTime":"2025-11-27T10:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.968847 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.968899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.968917 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.968943 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:15 crc kubenswrapper[4806]: I1127 10:22:15.968962 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:15Z","lastTransitionTime":"2025-11-27T10:22:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.071919 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.071971 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.071987 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.072011 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.072039 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:16Z","lastTransitionTime":"2025-11-27T10:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.116104 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.116403 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:16 crc kubenswrapper[4806]: E1127 10:22:16.116390 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.116501 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:16 crc kubenswrapper[4806]: E1127 10:22:16.116682 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:16 crc kubenswrapper[4806]: E1127 10:22:16.116916 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.141225 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.162439 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.174916 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.174983 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.174997 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.175023 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.175038 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:16Z","lastTransitionTime":"2025-11-27T10:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.183496 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.203160 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.246182 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"w:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1127 10:22:09.261320 6125 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.265093 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.279389 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.279714 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.279899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.280076 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.280206 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:16Z","lastTransitionTime":"2025-11-27T10:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.290803 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.307517 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.326969 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.350196 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.370290 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.383764 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.383805 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.383822 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.383846 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.383862 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:16Z","lastTransitionTime":"2025-11-27T10:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.393128 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.414538 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.434580 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.466663 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.482922 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.491098 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.491148 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.491162 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.491182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.491196 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:16Z","lastTransitionTime":"2025-11-27T10:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.497989 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.594324 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.594460 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.594483 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.594511 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.594565 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:16Z","lastTransitionTime":"2025-11-27T10:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.698810 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.698896 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.698912 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.698935 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.698987 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:16Z","lastTransitionTime":"2025-11-27T10:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.802750 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.802821 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.802844 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.802880 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.802905 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:16Z","lastTransitionTime":"2025-11-27T10:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.906964 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.907058 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.907127 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.907197 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:16 crc kubenswrapper[4806]: I1127 10:22:16.907221 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:16Z","lastTransitionTime":"2025-11-27T10:22:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.010504 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.010563 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.010582 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.010606 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.010623 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:17Z","lastTransitionTime":"2025-11-27T10:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.114446 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.114516 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.114539 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.114567 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.114589 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:17Z","lastTransitionTime":"2025-11-27T10:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.115995 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:17 crc kubenswrapper[4806]: E1127 10:22:17.116161 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.218897 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.218969 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.219019 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.219049 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.219068 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:17Z","lastTransitionTime":"2025-11-27T10:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.323018 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.323092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.323119 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.323150 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.323172 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:17Z","lastTransitionTime":"2025-11-27T10:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.426859 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.426926 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.426940 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.426968 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.427131 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:17Z","lastTransitionTime":"2025-11-27T10:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.529611 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.529657 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.529668 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.529683 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.529691 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:17Z","lastTransitionTime":"2025-11-27T10:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.632668 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.632715 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.632725 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.632749 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.632761 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:17Z","lastTransitionTime":"2025-11-27T10:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.735735 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.735794 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.735806 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.735831 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.735849 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:17Z","lastTransitionTime":"2025-11-27T10:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.839029 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.839073 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.839082 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.839100 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.839111 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:17Z","lastTransitionTime":"2025-11-27T10:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.947038 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.947140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.947154 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.947196 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:17 crc kubenswrapper[4806]: I1127 10:22:17.947216 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:17Z","lastTransitionTime":"2025-11-27T10:22:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.051758 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.051833 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.051846 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.051861 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.051872 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:18Z","lastTransitionTime":"2025-11-27T10:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.117614 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:18 crc kubenswrapper[4806]: E1127 10:22:18.117859 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.119665 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.119951 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:18 crc kubenswrapper[4806]: E1127 10:22:18.119953 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:18 crc kubenswrapper[4806]: E1127 10:22:18.120188 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.155183 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.155320 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.155351 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.155384 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.155408 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:18Z","lastTransitionTime":"2025-11-27T10:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.259649 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.260374 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.260415 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.260442 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.260457 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:18Z","lastTransitionTime":"2025-11-27T10:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.364392 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.364655 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.364771 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.364868 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.364958 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:18Z","lastTransitionTime":"2025-11-27T10:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.468341 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.468378 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.468389 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.468409 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.468421 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:18Z","lastTransitionTime":"2025-11-27T10:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.570614 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.570674 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.570684 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.570709 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.570724 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:18Z","lastTransitionTime":"2025-11-27T10:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.673162 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.673203 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.673214 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.673251 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.673263 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:18Z","lastTransitionTime":"2025-11-27T10:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.775989 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.776039 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.776048 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.776064 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.776075 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:18Z","lastTransitionTime":"2025-11-27T10:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.879333 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.879403 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.879421 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.879447 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.879463 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:18Z","lastTransitionTime":"2025-11-27T10:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.983143 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.983430 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.983454 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.983482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:18 crc kubenswrapper[4806]: I1127 10:22:18.983502 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:18Z","lastTransitionTime":"2025-11-27T10:22:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.087388 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.087460 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.087480 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.087508 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.087572 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:19Z","lastTransitionTime":"2025-11-27T10:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.116253 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:19 crc kubenswrapper[4806]: E1127 10:22:19.116466 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.191020 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.191080 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.191094 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.191122 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.191137 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:19Z","lastTransitionTime":"2025-11-27T10:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.295048 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.295119 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.295135 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.295160 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.295179 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:19Z","lastTransitionTime":"2025-11-27T10:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.398826 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.398902 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.398932 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.398962 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.398997 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:19Z","lastTransitionTime":"2025-11-27T10:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.502696 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.503160 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.503413 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.503601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.503772 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:19Z","lastTransitionTime":"2025-11-27T10:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.582448 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:19 crc kubenswrapper[4806]: E1127 10:22:19.582677 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:19 crc kubenswrapper[4806]: E1127 10:22:19.583006 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs podName:3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e nodeName:}" failed. No retries permitted until 2025-11-27 10:22:27.582981386 +0000 UTC m=+52.169572160 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs") pod "network-metrics-daemon-p2gpw" (UID: "3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.607109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.607187 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.607202 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.607226 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.607261 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:19Z","lastTransitionTime":"2025-11-27T10:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.710600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.710659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.710676 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.710696 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.710712 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:19Z","lastTransitionTime":"2025-11-27T10:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.813609 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.813659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.813676 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.813700 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.813715 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:19Z","lastTransitionTime":"2025-11-27T10:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.917557 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.918007 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.918094 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.918174 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:19 crc kubenswrapper[4806]: I1127 10:22:19.918260 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:19Z","lastTransitionTime":"2025-11-27T10:22:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.021008 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.021291 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.021424 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.021529 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.021622 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:20Z","lastTransitionTime":"2025-11-27T10:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.116191 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.116553 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:20 crc kubenswrapper[4806]: E1127 10:22:20.116734 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:20 crc kubenswrapper[4806]: E1127 10:22:20.116785 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.116838 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:20 crc kubenswrapper[4806]: E1127 10:22:20.117086 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.124308 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.124354 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.124366 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.124386 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.124401 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:20Z","lastTransitionTime":"2025-11-27T10:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.227017 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.227048 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.227057 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.227069 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.227077 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:20Z","lastTransitionTime":"2025-11-27T10:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.330873 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.330970 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.330989 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.331013 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.331031 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:20Z","lastTransitionTime":"2025-11-27T10:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.434863 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.434974 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.434992 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.435040 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.435052 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:20Z","lastTransitionTime":"2025-11-27T10:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.538376 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.538477 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.538536 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.538565 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.538584 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:20Z","lastTransitionTime":"2025-11-27T10:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.642077 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.642144 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.642165 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.642193 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.642211 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:20Z","lastTransitionTime":"2025-11-27T10:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.744785 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.744865 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.744885 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.744911 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.744933 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:20Z","lastTransitionTime":"2025-11-27T10:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.850140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.850527 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.850752 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.850963 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.851211 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:20Z","lastTransitionTime":"2025-11-27T10:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.954198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.954588 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.954885 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.955227 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:20 crc kubenswrapper[4806]: I1127 10:22:20.955720 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:20Z","lastTransitionTime":"2025-11-27T10:22:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.058856 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.058902 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.058918 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.058941 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.058956 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:21Z","lastTransitionTime":"2025-11-27T10:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.115861 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:21 crc kubenswrapper[4806]: E1127 10:22:21.116007 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.161619 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.161667 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.161678 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.161694 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.161705 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:21Z","lastTransitionTime":"2025-11-27T10:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.265133 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.265205 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.265220 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.265267 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.265284 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:21Z","lastTransitionTime":"2025-11-27T10:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.368078 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.368274 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.368302 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.368333 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.368355 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:21Z","lastTransitionTime":"2025-11-27T10:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.471437 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.471484 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.471496 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.471512 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.471522 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:21Z","lastTransitionTime":"2025-11-27T10:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.574455 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.574520 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.574537 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.574561 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.574579 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:21Z","lastTransitionTime":"2025-11-27T10:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.677640 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.677704 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.677721 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.677745 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.677763 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:21Z","lastTransitionTime":"2025-11-27T10:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.780616 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.781115 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.781489 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.781836 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.782016 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:21Z","lastTransitionTime":"2025-11-27T10:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.885560 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.886000 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.886086 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.886293 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.886421 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:21Z","lastTransitionTime":"2025-11-27T10:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.990211 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.990321 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.990344 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.990376 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:21 crc kubenswrapper[4806]: I1127 10:22:21.990411 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:21Z","lastTransitionTime":"2025-11-27T10:22:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.093563 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.093636 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.093660 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.093691 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.093721 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.116308 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.116504 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.116727 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:22 crc kubenswrapper[4806]: E1127 10:22:22.116729 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:22 crc kubenswrapper[4806]: E1127 10:22:22.116996 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:22 crc kubenswrapper[4806]: E1127 10:22:22.117061 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.197037 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.197466 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.197636 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.197772 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.197881 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.300657 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.300989 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.301288 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.301534 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.301633 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.405421 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.405463 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.405478 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.405501 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.405514 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.509861 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.509964 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.509981 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.510005 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.510017 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.586922 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.586992 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.587004 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.587025 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.587051 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: E1127 10:22:22.606181 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:22Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.611335 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.611394 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.611409 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.611430 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.611443 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: E1127 10:22:22.624724 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:22Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.628854 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.628948 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.629007 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.629088 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.629155 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: E1127 10:22:22.645521 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:22Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.649149 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.649198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.649212 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.649253 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.649267 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: E1127 10:22:22.662592 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:22Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.667327 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.667476 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.667542 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.667611 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.667680 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: E1127 10:22:22.681774 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:22Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:22 crc kubenswrapper[4806]: E1127 10:22:22.682112 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.683808 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.683857 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.683870 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.683892 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.683905 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.786987 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.787047 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.787060 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.787084 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.787103 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.890370 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.890430 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.890450 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.890471 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.890486 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.993522 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.993579 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.993596 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.993621 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:22 crc kubenswrapper[4806]: I1127 10:22:22.993638 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:22Z","lastTransitionTime":"2025-11-27T10:22:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.096145 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.096220 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.096277 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.096302 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.096322 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:23Z","lastTransitionTime":"2025-11-27T10:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.116685 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:23 crc kubenswrapper[4806]: E1127 10:22:23.117292 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.117711 4806 scope.go:117] "RemoveContainer" containerID="29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.200362 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.200417 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.200439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.200467 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.200488 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:23Z","lastTransitionTime":"2025-11-27T10:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.411152 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.411224 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.411334 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.411379 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.411422 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:23Z","lastTransitionTime":"2025-11-27T10:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.514320 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.514365 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.514381 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.514405 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.514424 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:23Z","lastTransitionTime":"2025-11-27T10:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.619203 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.619701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.619722 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.619753 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.619772 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:23Z","lastTransitionTime":"2025-11-27T10:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.723736 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.723793 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.723804 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.723828 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.723839 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:23Z","lastTransitionTime":"2025-11-27T10:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.827170 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.827223 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.827261 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.827283 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.827296 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:23Z","lastTransitionTime":"2025-11-27T10:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.929625 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.929662 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.929670 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.929684 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:23 crc kubenswrapper[4806]: I1127 10:22:23.929692 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:23Z","lastTransitionTime":"2025-11-27T10:22:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.032383 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.032451 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.032467 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.032497 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.032511 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:24Z","lastTransitionTime":"2025-11-27T10:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.116047 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:24 crc kubenswrapper[4806]: E1127 10:22:24.116168 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.116178 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:24 crc kubenswrapper[4806]: E1127 10:22:24.116247 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.116504 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:24 crc kubenswrapper[4806]: E1127 10:22:24.116577 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.135070 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.135097 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.135106 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.135128 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.135137 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:24Z","lastTransitionTime":"2025-11-27T10:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.237789 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.237824 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.237835 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.237852 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.237863 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:24Z","lastTransitionTime":"2025-11-27T10:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.340550 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.340605 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.340622 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.340643 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.340661 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:24Z","lastTransitionTime":"2025-11-27T10:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.443182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.443969 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.444045 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.444117 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.444177 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:24Z","lastTransitionTime":"2025-11-27T10:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.459918 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/2.log" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.460457 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/1.log" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.463948 4806 generic.go:334] "Generic (PLEG): container finished" podID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerID="eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993" exitCode=1 Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.463987 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993"} Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.464026 4806 scope.go:117] "RemoveContainer" containerID="29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.464826 4806 scope.go:117] "RemoveContainer" containerID="eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993" Nov 27 10:22:24 crc kubenswrapper[4806]: E1127 10:22:24.465061 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.478609 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.488772 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.505903 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"w:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1127 10:22:09.261320 6125 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:24Z\\\",\\\"message\\\":\\\":[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:24.421527 6328 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"d4efc4a8-c514-4a6b-901c-2953978b50d3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.518656 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.536439 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.547000 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.547102 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.547165 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.547190 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.547208 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:24Z","lastTransitionTime":"2025-11-27T10:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.548641 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.562035 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.576634 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.589140 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.602711 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.618844 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.630776 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.649802 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.650105 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.650262 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.650441 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.650594 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:24Z","lastTransitionTime":"2025-11-27T10:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.653710 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.667617 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.682289 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.697914 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.713174 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.753785 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.754031 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.754172 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.754341 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.754509 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:24Z","lastTransitionTime":"2025-11-27T10:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.857527 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.857593 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.857615 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.857645 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.857668 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:24Z","lastTransitionTime":"2025-11-27T10:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.961284 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.961597 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.961734 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.961840 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:24 crc kubenswrapper[4806]: I1127 10:22:24.961945 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:24Z","lastTransitionTime":"2025-11-27T10:22:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.065944 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.066006 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.066030 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.066060 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.066083 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:25Z","lastTransitionTime":"2025-11-27T10:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.116707 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:25 crc kubenswrapper[4806]: E1127 10:22:25.116941 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.168274 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.168353 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.168375 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.168401 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.168419 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:25Z","lastTransitionTime":"2025-11-27T10:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.271502 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.271550 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.271566 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.271588 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.271605 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:25Z","lastTransitionTime":"2025-11-27T10:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.374592 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.374654 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.374675 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.374701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.374722 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:25Z","lastTransitionTime":"2025-11-27T10:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.470203 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/2.log" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.481925 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.481980 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.482001 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.482027 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.482046 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:25Z","lastTransitionTime":"2025-11-27T10:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.585682 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.586016 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.586184 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.586362 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.586552 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:25Z","lastTransitionTime":"2025-11-27T10:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.647646 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.660687 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.667521 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.683371 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.689946 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.690371 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.690565 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.690752 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.690992 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:25Z","lastTransitionTime":"2025-11-27T10:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.706968 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"w:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1127 10:22:09.261320 6125 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:24Z\\\",\\\"message\\\":\\\":[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:24.421527 6328 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"d4efc4a8-c514-4a6b-901c-2953978b50d3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.721821 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.735723 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.751449 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.768614 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.785020 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.793992 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.794033 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.794048 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.794069 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.794089 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:25Z","lastTransitionTime":"2025-11-27T10:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.803364 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.818284 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.831506 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.860082 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.875200 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.891683 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.896364 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.896408 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.896423 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.896444 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.896459 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:25Z","lastTransitionTime":"2025-11-27T10:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.906418 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.923096 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.943534 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:25Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.998512 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.998557 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.998572 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.998590 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:25 crc kubenswrapper[4806]: I1127 10:22:25.998604 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:25Z","lastTransitionTime":"2025-11-27T10:22:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.101546 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.101599 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.101613 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.101633 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.101646 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:26Z","lastTransitionTime":"2025-11-27T10:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.116453 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.116509 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.116547 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:26 crc kubenswrapper[4806]: E1127 10:22:26.116701 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:26 crc kubenswrapper[4806]: E1127 10:22:26.117199 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:26 crc kubenswrapper[4806]: E1127 10:22:26.117318 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.135775 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.149111 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.179205 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.195393 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.204352 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.204395 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.204413 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.204432 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.204446 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:26Z","lastTransitionTime":"2025-11-27T10:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.213051 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.234401 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.259609 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.278829 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.297288 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.309539 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.309626 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.309655 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.309685 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.309709 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:26Z","lastTransitionTime":"2025-11-27T10:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.314132 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.337389 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"w:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1127 10:22:09.261320 6125 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:24Z\\\",\\\"message\\\":\\\":[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:24.421527 6328 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"d4efc4a8-c514-4a6b-901c-2953978b50d3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.351486 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.363070 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.378169 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93396449-9bb1-4cfc-bba4-5c00c92cdf6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6784c0e140b36ee0c0a9135631f867702e99bd54be29abcea1bd7d97f5c533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2d5c2fdd37029ccfa325eb3a2f2fa6d1bcb65b3847dd4997e2cc12798603c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe72cadf6271151c95f0efb2cd1e154c78d1398d9f104c57acc255ecc99738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.396550 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.412745 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.412805 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.412824 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.412848 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.412866 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:26Z","lastTransitionTime":"2025-11-27T10:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.418637 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.433768 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.459970 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:26Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.515825 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.515861 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.515871 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.515885 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.515896 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:26Z","lastTransitionTime":"2025-11-27T10:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.618966 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.619017 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.619038 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.619066 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.619086 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:26Z","lastTransitionTime":"2025-11-27T10:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.722267 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.722337 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.722355 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.722382 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.722401 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:26Z","lastTransitionTime":"2025-11-27T10:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.827425 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.827697 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.827813 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.827916 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.828009 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:26Z","lastTransitionTime":"2025-11-27T10:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.930419 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.930470 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.930482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.930501 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:26 crc kubenswrapper[4806]: I1127 10:22:26.930512 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:26Z","lastTransitionTime":"2025-11-27T10:22:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.033629 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.033671 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.033679 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.033692 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.033700 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:27Z","lastTransitionTime":"2025-11-27T10:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.116216 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.116479 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.136459 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.136487 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.136498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.136512 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.136522 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:27Z","lastTransitionTime":"2025-11-27T10:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.240348 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.240446 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.240475 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.240512 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.240551 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:27Z","lastTransitionTime":"2025-11-27T10:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.343832 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.344109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.344242 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.344365 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.344492 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:27Z","lastTransitionTime":"2025-11-27T10:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.447532 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.447601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.447620 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.447646 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.447665 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:27Z","lastTransitionTime":"2025-11-27T10:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.552492 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.552571 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.552580 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.552594 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.552604 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:27Z","lastTransitionTime":"2025-11-27T10:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.655113 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.655378 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.655485 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs podName:3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e nodeName:}" failed. No retries permitted until 2025-11-27 10:22:43.655456931 +0000 UTC m=+68.242047725 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs") pod "network-metrics-daemon-p2gpw" (UID: "3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.657944 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.658153 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.658364 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.658588 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.658787 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:27Z","lastTransitionTime":"2025-11-27T10:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.756047 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.756394 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.756524 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:59.756492662 +0000 UTC m=+84.343083466 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.761898 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.761930 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.761940 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.761954 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.761963 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:27Z","lastTransitionTime":"2025-11-27T10:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.857502 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.857648 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.857695 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.857729 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.857809 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:22:59.857743549 +0000 UTC m=+84.444334353 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.857835 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.857889 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.857921 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.857815 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.857942 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.857901 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:59.857880533 +0000 UTC m=+84.444471327 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.857974 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.857997 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.857997 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:59.857979196 +0000 UTC m=+84.444570000 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:27 crc kubenswrapper[4806]: E1127 10:22:27.858088 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 10:22:59.858071678 +0000 UTC m=+84.444662482 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.865482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.865540 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.865559 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.865582 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.865599 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:27Z","lastTransitionTime":"2025-11-27T10:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.968327 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.968403 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.968420 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.968443 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:27 crc kubenswrapper[4806]: I1127 10:22:27.968461 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:27Z","lastTransitionTime":"2025-11-27T10:22:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.071804 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.071859 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.071876 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.071898 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.071917 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:28Z","lastTransitionTime":"2025-11-27T10:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.116553 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.116596 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.116569 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:28 crc kubenswrapper[4806]: E1127 10:22:28.116723 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:28 crc kubenswrapper[4806]: E1127 10:22:28.116837 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:28 crc kubenswrapper[4806]: E1127 10:22:28.116924 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.174758 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.174817 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.174834 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.174856 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.174872 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:28Z","lastTransitionTime":"2025-11-27T10:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.277829 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.277897 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.277920 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.277953 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.277976 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:28Z","lastTransitionTime":"2025-11-27T10:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.380845 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.380893 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.380909 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.380933 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.380950 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:28Z","lastTransitionTime":"2025-11-27T10:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.484008 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.484061 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.484079 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.484103 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.484121 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:28Z","lastTransitionTime":"2025-11-27T10:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.586387 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.586446 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.586469 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.586498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.586515 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:28Z","lastTransitionTime":"2025-11-27T10:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.689634 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.689672 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.689685 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.689704 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.689720 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:28Z","lastTransitionTime":"2025-11-27T10:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.792614 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.792694 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.792716 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.792748 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.792770 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:28Z","lastTransitionTime":"2025-11-27T10:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.895082 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.895143 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.895159 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.895182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.895199 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:28Z","lastTransitionTime":"2025-11-27T10:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.997887 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.997951 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.997969 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.997990 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:28 crc kubenswrapper[4806]: I1127 10:22:28.998007 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:28Z","lastTransitionTime":"2025-11-27T10:22:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.100806 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.100871 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.100890 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.100915 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.100932 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:29Z","lastTransitionTime":"2025-11-27T10:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.116179 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:29 crc kubenswrapper[4806]: E1127 10:22:29.116611 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.203193 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.203247 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.203263 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.203280 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.203291 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:29Z","lastTransitionTime":"2025-11-27T10:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.305416 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.305475 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.305493 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.305519 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.305536 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:29Z","lastTransitionTime":"2025-11-27T10:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.409491 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.409571 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.409596 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.409628 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.409652 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:29Z","lastTransitionTime":"2025-11-27T10:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.512600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.512765 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.512844 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.512932 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.513037 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:29Z","lastTransitionTime":"2025-11-27T10:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.616320 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.616956 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.617040 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.617153 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.617260 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:29Z","lastTransitionTime":"2025-11-27T10:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.720255 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.720315 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.720331 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.720356 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.720370 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:29Z","lastTransitionTime":"2025-11-27T10:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.823707 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.823768 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.823785 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.823808 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.823827 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:29Z","lastTransitionTime":"2025-11-27T10:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.926626 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.926699 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.926718 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.926746 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:29 crc kubenswrapper[4806]: I1127 10:22:29.926772 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:29Z","lastTransitionTime":"2025-11-27T10:22:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.029482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.029561 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.029578 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.029601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.029617 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:30Z","lastTransitionTime":"2025-11-27T10:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.116662 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.116701 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:30 crc kubenswrapper[4806]: E1127 10:22:30.116889 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.117097 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:30 crc kubenswrapper[4806]: E1127 10:22:30.117225 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:30 crc kubenswrapper[4806]: E1127 10:22:30.117485 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.131943 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.131998 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.132016 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.132043 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.132066 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:30Z","lastTransitionTime":"2025-11-27T10:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.235180 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.235294 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.235319 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.235350 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.235375 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:30Z","lastTransitionTime":"2025-11-27T10:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.337555 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.337594 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.337604 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.337620 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.337629 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:30Z","lastTransitionTime":"2025-11-27T10:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.441310 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.441383 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.441409 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.441440 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.441458 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:30Z","lastTransitionTime":"2025-11-27T10:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.543980 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.544083 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.544107 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.544136 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.544156 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:30Z","lastTransitionTime":"2025-11-27T10:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.651887 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.651922 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.651933 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.651948 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.651958 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:30Z","lastTransitionTime":"2025-11-27T10:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.755062 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.755553 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.755760 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.755969 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.756121 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:30Z","lastTransitionTime":"2025-11-27T10:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.860187 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.860779 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.861150 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.861712 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.861939 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:30Z","lastTransitionTime":"2025-11-27T10:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.965018 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.965088 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.965111 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.965139 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:30 crc kubenswrapper[4806]: I1127 10:22:30.965159 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:30Z","lastTransitionTime":"2025-11-27T10:22:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.068857 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.068919 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.068940 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.068968 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.068993 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:31Z","lastTransitionTime":"2025-11-27T10:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.116294 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:31 crc kubenswrapper[4806]: E1127 10:22:31.116876 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.172122 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.172192 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.172214 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.172309 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.172334 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:31Z","lastTransitionTime":"2025-11-27T10:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.275820 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.275898 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.275917 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.275945 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.275963 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:31Z","lastTransitionTime":"2025-11-27T10:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.379615 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.379677 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.379697 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.379733 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.379751 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:31Z","lastTransitionTime":"2025-11-27T10:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.482654 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.482723 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.482741 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.482767 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.482785 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:31Z","lastTransitionTime":"2025-11-27T10:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.585907 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.585980 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.586001 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.586032 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.586056 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:31Z","lastTransitionTime":"2025-11-27T10:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.688679 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.688749 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.688768 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.688792 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.688810 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:31Z","lastTransitionTime":"2025-11-27T10:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.791940 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.792326 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.792489 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.792641 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.792784 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:31Z","lastTransitionTime":"2025-11-27T10:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.895662 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.895695 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.895702 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.895716 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.895724 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:31Z","lastTransitionTime":"2025-11-27T10:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.999269 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.999302 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.999313 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.999329 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:31 crc kubenswrapper[4806]: I1127 10:22:31.999339 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:31Z","lastTransitionTime":"2025-11-27T10:22:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.102275 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.102336 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.102353 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.102376 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.102394 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:32Z","lastTransitionTime":"2025-11-27T10:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.115796 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.115828 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:32 crc kubenswrapper[4806]: E1127 10:22:32.115995 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.116028 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:32 crc kubenswrapper[4806]: E1127 10:22:32.116122 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:32 crc kubenswrapper[4806]: E1127 10:22:32.116313 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.205303 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.205360 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.205381 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.205407 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.205427 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:32Z","lastTransitionTime":"2025-11-27T10:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.308125 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.308188 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.308209 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.308270 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.308296 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:32Z","lastTransitionTime":"2025-11-27T10:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.410985 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.411050 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.411071 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.411102 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.411119 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:32Z","lastTransitionTime":"2025-11-27T10:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.513577 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.513651 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.513673 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.513696 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.513714 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:32Z","lastTransitionTime":"2025-11-27T10:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.617008 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.617084 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.617111 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.617141 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.617164 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:32Z","lastTransitionTime":"2025-11-27T10:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.720795 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.720856 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.720874 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.720900 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.720918 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:32Z","lastTransitionTime":"2025-11-27T10:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.824380 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.824447 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.824464 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.824489 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.824508 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:32Z","lastTransitionTime":"2025-11-27T10:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.918964 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.919049 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.919069 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.919091 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.919108 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:32Z","lastTransitionTime":"2025-11-27T10:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:32 crc kubenswrapper[4806]: E1127 10:22:32.939948 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:32Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.945536 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.945607 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.945630 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.945654 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.945670 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:32Z","lastTransitionTime":"2025-11-27T10:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:32 crc kubenswrapper[4806]: E1127 10:22:32.965776 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:32Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.972000 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.972069 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.972093 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.972125 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:32 crc kubenswrapper[4806]: I1127 10:22:32.972147 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:32Z","lastTransitionTime":"2025-11-27T10:22:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:32 crc kubenswrapper[4806]: E1127 10:22:32.993316 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:32Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.005359 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.005506 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.005529 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.005609 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.005637 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:33Z","lastTransitionTime":"2025-11-27T10:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:33 crc kubenswrapper[4806]: E1127 10:22:33.037388 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:33Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.041397 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.041431 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.041444 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.041461 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.041475 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:33Z","lastTransitionTime":"2025-11-27T10:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:33 crc kubenswrapper[4806]: E1127 10:22:33.057620 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:33Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:33 crc kubenswrapper[4806]: E1127 10:22:33.058018 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.060464 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.060540 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.060565 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.060596 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.060625 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:33Z","lastTransitionTime":"2025-11-27T10:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.115994 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:33 crc kubenswrapper[4806]: E1127 10:22:33.116207 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.163073 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.163143 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.163170 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.163199 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.163217 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:33Z","lastTransitionTime":"2025-11-27T10:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.265835 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.265904 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.265928 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.265957 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.265982 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:33Z","lastTransitionTime":"2025-11-27T10:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.369397 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.369444 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.369460 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.369482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.369499 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:33Z","lastTransitionTime":"2025-11-27T10:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.473139 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.473207 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.473227 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.473284 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.473303 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:33Z","lastTransitionTime":"2025-11-27T10:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.577000 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.577041 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.577052 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.577067 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.577081 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:33Z","lastTransitionTime":"2025-11-27T10:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.681221 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.681323 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.681341 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.681371 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.681388 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:33Z","lastTransitionTime":"2025-11-27T10:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.785010 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.785100 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.785123 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.785154 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.785177 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:33Z","lastTransitionTime":"2025-11-27T10:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.889353 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.889433 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.889451 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.889477 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.889492 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:33Z","lastTransitionTime":"2025-11-27T10:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.994300 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.994365 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.994377 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.994400 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:33 crc kubenswrapper[4806]: I1127 10:22:33.994417 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:33Z","lastTransitionTime":"2025-11-27T10:22:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.097499 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.097557 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.097570 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.097589 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.097603 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:34Z","lastTransitionTime":"2025-11-27T10:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.116092 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.116140 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.116266 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:34 crc kubenswrapper[4806]: E1127 10:22:34.116326 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:34 crc kubenswrapper[4806]: E1127 10:22:34.116502 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:34 crc kubenswrapper[4806]: E1127 10:22:34.116591 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.200783 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.200839 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.200853 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.200876 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.200888 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:34Z","lastTransitionTime":"2025-11-27T10:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.303952 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.304009 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.304023 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.304081 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.304099 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:34Z","lastTransitionTime":"2025-11-27T10:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.407461 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.407514 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.407526 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.407544 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.407554 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:34Z","lastTransitionTime":"2025-11-27T10:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.510203 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.510267 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.510296 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.510314 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.510324 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:34Z","lastTransitionTime":"2025-11-27T10:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.613452 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.613512 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.613526 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.613548 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.613562 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:34Z","lastTransitionTime":"2025-11-27T10:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.716678 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.716749 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.716768 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.716795 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.716824 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:34Z","lastTransitionTime":"2025-11-27T10:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.818786 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.818841 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.818852 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.818869 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.818881 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:34Z","lastTransitionTime":"2025-11-27T10:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.921610 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.921681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.921703 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.921737 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:34 crc kubenswrapper[4806]: I1127 10:22:34.921762 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:34Z","lastTransitionTime":"2025-11-27T10:22:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.028610 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.028698 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.028724 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.028754 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.028787 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:35Z","lastTransitionTime":"2025-11-27T10:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.115771 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:35 crc kubenswrapper[4806]: E1127 10:22:35.115910 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.131004 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.131047 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.131058 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.131073 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.131083 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:35Z","lastTransitionTime":"2025-11-27T10:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.234645 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.234713 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.234729 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.234752 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.234769 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:35Z","lastTransitionTime":"2025-11-27T10:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.338994 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.339089 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.339108 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.339603 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.339832 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:35Z","lastTransitionTime":"2025-11-27T10:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.445187 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.445281 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.445305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.445372 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.445393 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:35Z","lastTransitionTime":"2025-11-27T10:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.548759 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.548811 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.548822 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.548842 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.548854 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:35Z","lastTransitionTime":"2025-11-27T10:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.652510 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.652576 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.652594 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.652621 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.652637 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:35Z","lastTransitionTime":"2025-11-27T10:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.755753 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.755797 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.755808 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.755826 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.755839 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:35Z","lastTransitionTime":"2025-11-27T10:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.859528 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.859612 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.859626 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.859662 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.859677 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:35Z","lastTransitionTime":"2025-11-27T10:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.963678 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.963730 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.963744 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.963766 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:35 crc kubenswrapper[4806]: I1127 10:22:35.963779 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:35Z","lastTransitionTime":"2025-11-27T10:22:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.066466 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.066517 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.066530 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.066551 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.066566 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:36Z","lastTransitionTime":"2025-11-27T10:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.117078 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:36 crc kubenswrapper[4806]: E1127 10:22:36.117289 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.117875 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.117912 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:36 crc kubenswrapper[4806]: E1127 10:22:36.117961 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:36 crc kubenswrapper[4806]: E1127 10:22:36.118084 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.140066 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.156855 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.170156 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.170214 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.170261 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.170289 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.170307 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:36Z","lastTransitionTime":"2025-11-27T10:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.173365 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.190088 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.218317 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.234442 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.252854 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.268954 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.273462 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.273501 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.273522 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.273547 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.273560 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:36Z","lastTransitionTime":"2025-11-27T10:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.283100 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.296710 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.308135 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.330152 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29961af0c7f7ac21b7b651bf2229038efed64e58147e305975a97417a81dbb9a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"message\\\":\\\"w:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1127 10:22:09.261320 6125 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:24Z\\\",\\\"message\\\":\\\":[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:24.421527 6328 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"d4efc4a8-c514-4a6b-901c-2953978b50d3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.342664 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.357581 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.370504 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.376873 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.376923 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.376933 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.376951 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.376964 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:36Z","lastTransitionTime":"2025-11-27T10:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.391755 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93396449-9bb1-4cfc-bba4-5c00c92cdf6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6784c0e140b36ee0c0a9135631f867702e99bd54be29abcea1bd7d97f5c533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2d5c2fdd37029ccfa325eb3a2f2fa6d1bcb65b3847dd4997e2cc12798603c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe72cadf6271151c95f0efb2cd1e154c78d1398d9f104c57acc255ecc99738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.406101 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.422332 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:36Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.479993 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.480039 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.480053 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.480074 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.480087 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:36Z","lastTransitionTime":"2025-11-27T10:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.583023 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.583076 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.583085 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.583105 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.583116 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:36Z","lastTransitionTime":"2025-11-27T10:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.685835 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.685915 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.685933 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.685956 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.685976 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:36Z","lastTransitionTime":"2025-11-27T10:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.789070 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.789515 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.789622 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.789719 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.789821 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:36Z","lastTransitionTime":"2025-11-27T10:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.893453 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.893512 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.893531 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.893553 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.893569 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:36Z","lastTransitionTime":"2025-11-27T10:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.996458 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.996506 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.996516 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.996532 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:36 crc kubenswrapper[4806]: I1127 10:22:36.996542 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:36Z","lastTransitionTime":"2025-11-27T10:22:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.100113 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.100148 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.100157 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.100173 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.100182 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:37Z","lastTransitionTime":"2025-11-27T10:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.116126 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.116974 4806 scope.go:117] "RemoveContainer" containerID="eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993" Nov 27 10:22:37 crc kubenswrapper[4806]: E1127 10:22:37.117098 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:37 crc kubenswrapper[4806]: E1127 10:22:37.117385 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.134998 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.148801 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.162197 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.173818 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.193813 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:24Z\\\",\\\"message\\\":\\\":[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:24.421527 6328 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"d4efc4a8-c514-4a6b-901c-2953978b50d3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.203290 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.203323 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.203334 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.203349 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.203362 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:37Z","lastTransitionTime":"2025-11-27T10:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.209658 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.224303 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93396449-9bb1-4cfc-bba4-5c00c92cdf6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6784c0e140b36ee0c0a9135631f867702e99bd54be29abcea1bd7d97f5c533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2d5c2fdd37029ccfa325eb3a2f2fa6d1bcb65b3847dd4997e2cc12798603c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe72cadf6271151c95f0efb2cd1e154c78d1398d9f104c57acc255ecc99738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.242329 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.259708 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.276453 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.296976 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.308000 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.308034 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.308044 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.308056 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.308067 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:37Z","lastTransitionTime":"2025-11-27T10:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.311571 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.324726 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.349666 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.369338 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.383989 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.401438 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.412050 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.412372 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.412462 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.412551 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.412644 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:37Z","lastTransitionTime":"2025-11-27T10:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.418017 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:37Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.515675 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.515729 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.515744 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.515763 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.515775 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:37Z","lastTransitionTime":"2025-11-27T10:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.618524 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.618597 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.618613 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.618640 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.618659 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:37Z","lastTransitionTime":"2025-11-27T10:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.722731 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.722791 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.722803 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.722827 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.722838 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:37Z","lastTransitionTime":"2025-11-27T10:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.825029 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.825075 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.825109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.825126 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.825135 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:37Z","lastTransitionTime":"2025-11-27T10:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.928276 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.928338 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.928353 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.928376 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:37 crc kubenswrapper[4806]: I1127 10:22:37.928394 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:37Z","lastTransitionTime":"2025-11-27T10:22:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.031164 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.031252 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.031266 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.031287 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.031302 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:38Z","lastTransitionTime":"2025-11-27T10:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.116745 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:38 crc kubenswrapper[4806]: E1127 10:22:38.116958 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.116770 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.116770 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:38 crc kubenswrapper[4806]: E1127 10:22:38.117181 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:38 crc kubenswrapper[4806]: E1127 10:22:38.117312 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.133705 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.133741 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.133750 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.133768 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.133782 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:38Z","lastTransitionTime":"2025-11-27T10:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.236381 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.236688 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.236781 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.236881 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.236979 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:38Z","lastTransitionTime":"2025-11-27T10:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.341887 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.342479 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.342685 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.342829 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.342971 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:38Z","lastTransitionTime":"2025-11-27T10:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.446158 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.446219 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.446274 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.446303 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.446364 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:38Z","lastTransitionTime":"2025-11-27T10:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.549666 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.549698 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.549706 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.549717 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.549726 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:38Z","lastTransitionTime":"2025-11-27T10:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.660501 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.660549 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.660566 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.660587 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.660604 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:38Z","lastTransitionTime":"2025-11-27T10:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.763177 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.763222 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.763268 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.763288 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.763304 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:38Z","lastTransitionTime":"2025-11-27T10:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.865601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.865643 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.865659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.865681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.865698 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:38Z","lastTransitionTime":"2025-11-27T10:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.969107 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.969146 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.969162 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.969182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:38 crc kubenswrapper[4806]: I1127 10:22:38.969197 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:38Z","lastTransitionTime":"2025-11-27T10:22:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.072457 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.072523 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.072547 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.072575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.072599 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:39Z","lastTransitionTime":"2025-11-27T10:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.116728 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:39 crc kubenswrapper[4806]: E1127 10:22:39.116991 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.176504 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.176597 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.176622 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.176655 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.176678 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:39Z","lastTransitionTime":"2025-11-27T10:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.279908 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.280261 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.280372 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.280474 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.280587 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:39Z","lastTransitionTime":"2025-11-27T10:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.383552 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.383600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.383615 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.383630 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.383643 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:39Z","lastTransitionTime":"2025-11-27T10:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.486034 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.486078 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.486094 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.486113 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.486128 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:39Z","lastTransitionTime":"2025-11-27T10:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.588101 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.588463 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.588594 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.588752 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.588878 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:39Z","lastTransitionTime":"2025-11-27T10:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.690845 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.691165 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.691338 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.691492 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.691642 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:39Z","lastTransitionTime":"2025-11-27T10:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.794357 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.794406 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.794415 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.794437 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.794449 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:39Z","lastTransitionTime":"2025-11-27T10:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.897688 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.897746 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.897760 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.897785 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:39 crc kubenswrapper[4806]: I1127 10:22:39.897800 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:39Z","lastTransitionTime":"2025-11-27T10:22:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.000834 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.001687 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.001789 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.001895 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.001999 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:40Z","lastTransitionTime":"2025-11-27T10:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.104498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.104538 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.104551 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.104564 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.104576 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:40Z","lastTransitionTime":"2025-11-27T10:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.116024 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.116024 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:40 crc kubenswrapper[4806]: E1127 10:22:40.116300 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:40 crc kubenswrapper[4806]: E1127 10:22:40.116333 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.116037 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:40 crc kubenswrapper[4806]: E1127 10:22:40.116413 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.207917 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.207991 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.208004 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.208035 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.208263 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:40Z","lastTransitionTime":"2025-11-27T10:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.311276 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.311325 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.311342 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.311365 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.311382 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:40Z","lastTransitionTime":"2025-11-27T10:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.414303 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.414364 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.414396 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.414424 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.414441 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:40Z","lastTransitionTime":"2025-11-27T10:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.517537 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.517594 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.517617 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.517647 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.517670 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:40Z","lastTransitionTime":"2025-11-27T10:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.620778 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.620832 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.620845 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.620865 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.620878 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:40Z","lastTransitionTime":"2025-11-27T10:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.724299 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.724351 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.724364 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.724386 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.724397 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:40Z","lastTransitionTime":"2025-11-27T10:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.826943 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.826982 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.826993 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.827011 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.827021 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:40Z","lastTransitionTime":"2025-11-27T10:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.930538 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.930575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.930591 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.930609 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:40 crc kubenswrapper[4806]: I1127 10:22:40.930621 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:40Z","lastTransitionTime":"2025-11-27T10:22:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.033405 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.033441 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.033453 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.033470 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.033481 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:41Z","lastTransitionTime":"2025-11-27T10:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.116342 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:41 crc kubenswrapper[4806]: E1127 10:22:41.116651 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.136270 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.136341 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.136358 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.136388 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.136406 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:41Z","lastTransitionTime":"2025-11-27T10:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.238336 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.239465 4806 scope.go:117] "RemoveContainer" containerID="eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.239508 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.239566 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.239581 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.239601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.239616 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:41Z","lastTransitionTime":"2025-11-27T10:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:41 crc kubenswrapper[4806]: E1127 10:22:41.239658 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.343131 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.343173 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.343185 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.343200 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.343211 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:41Z","lastTransitionTime":"2025-11-27T10:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.446678 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.446727 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.446743 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.446766 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.446781 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:41Z","lastTransitionTime":"2025-11-27T10:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.549548 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.549586 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.549600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.549619 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.549631 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:41Z","lastTransitionTime":"2025-11-27T10:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.652534 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.652585 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.652600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.652622 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.652636 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:41Z","lastTransitionTime":"2025-11-27T10:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.755800 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.755853 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.755866 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.755887 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.755899 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:41Z","lastTransitionTime":"2025-11-27T10:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.858678 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.859091 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.859289 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.859573 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.859742 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:41Z","lastTransitionTime":"2025-11-27T10:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.962815 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.963292 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.963370 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.963466 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:41 crc kubenswrapper[4806]: I1127 10:22:41.963544 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:41Z","lastTransitionTime":"2025-11-27T10:22:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.067560 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.067621 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.067634 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.067659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.067672 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:42Z","lastTransitionTime":"2025-11-27T10:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.115804 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.115884 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.115823 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:42 crc kubenswrapper[4806]: E1127 10:22:42.116193 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:42 crc kubenswrapper[4806]: E1127 10:22:42.116366 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:42 crc kubenswrapper[4806]: E1127 10:22:42.116482 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.170650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.170700 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.170713 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.170734 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.170748 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:42Z","lastTransitionTime":"2025-11-27T10:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.274362 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.274427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.274443 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.274469 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.274489 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:42Z","lastTransitionTime":"2025-11-27T10:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.377729 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.378272 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.378387 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.378488 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.378554 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:42Z","lastTransitionTime":"2025-11-27T10:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.481243 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.481285 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.481298 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.481316 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.481328 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:42Z","lastTransitionTime":"2025-11-27T10:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.584446 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.584483 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.584496 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.584513 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.584524 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:42Z","lastTransitionTime":"2025-11-27T10:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.687590 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.687659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.687674 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.687697 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.687709 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:42Z","lastTransitionTime":"2025-11-27T10:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.790012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.790360 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.790467 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.790544 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.790604 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:42Z","lastTransitionTime":"2025-11-27T10:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.892924 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.892984 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.892998 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.893023 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.893039 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:42Z","lastTransitionTime":"2025-11-27T10:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.995862 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.995916 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.995928 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.995946 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:42 crc kubenswrapper[4806]: I1127 10:22:42.995957 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:42Z","lastTransitionTime":"2025-11-27T10:22:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.100801 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.100855 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.100866 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.100891 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.100911 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.115721 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:43 crc kubenswrapper[4806]: E1127 10:22:43.115903 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.203803 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.203851 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.203861 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.203877 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.203888 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.307686 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.307729 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.307739 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.307756 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.307766 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.328754 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.329262 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.329368 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.329479 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.329573 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: E1127 10:22:43.346150 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:43Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.351030 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.351086 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.351096 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.351118 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.351129 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: E1127 10:22:43.364842 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:43Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.370476 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.370527 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.370540 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.370564 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.370578 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: E1127 10:22:43.384597 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:43Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.390033 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.390106 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.390118 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.390148 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.390178 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: E1127 10:22:43.404060 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:43Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.409862 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.409919 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.409934 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.409956 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.409980 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: E1127 10:22:43.424011 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:43Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:43 crc kubenswrapper[4806]: E1127 10:22:43.424192 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.426434 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.426477 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.426492 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.426512 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.426528 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.529588 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.529644 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.529665 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.529690 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.529704 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.632171 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.632211 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.632221 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.632251 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.632266 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.735498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.735545 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.735556 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.735575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.735586 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.740008 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:43 crc kubenswrapper[4806]: E1127 10:22:43.740163 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:43 crc kubenswrapper[4806]: E1127 10:22:43.740224 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs podName:3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e nodeName:}" failed. No retries permitted until 2025-11-27 10:23:15.740205826 +0000 UTC m=+100.326796590 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs") pod "network-metrics-daemon-p2gpw" (UID: "3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.838471 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.838536 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.838556 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.838581 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.838598 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.941957 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.942039 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.942054 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.942077 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:43 crc kubenswrapper[4806]: I1127 10:22:43.942092 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:43Z","lastTransitionTime":"2025-11-27T10:22:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.044986 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.045036 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.045049 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.045068 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.045081 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:44Z","lastTransitionTime":"2025-11-27T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.116010 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.116080 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.116121 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:44 crc kubenswrapper[4806]: E1127 10:22:44.116347 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:44 crc kubenswrapper[4806]: E1127 10:22:44.116474 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:44 crc kubenswrapper[4806]: E1127 10:22:44.116569 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.148077 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.148132 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.148144 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.148158 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.148167 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:44Z","lastTransitionTime":"2025-11-27T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.250401 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.250438 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.250451 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.250465 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.250475 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:44Z","lastTransitionTime":"2025-11-27T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.352499 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.352526 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.352534 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.352547 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.352555 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:44Z","lastTransitionTime":"2025-11-27T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.454539 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.454564 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.454571 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.454583 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.454591 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:44Z","lastTransitionTime":"2025-11-27T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.556478 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.556526 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.556538 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.556556 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.556569 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:44Z","lastTransitionTime":"2025-11-27T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.664354 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.664407 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.664420 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.664436 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.664523 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:44Z","lastTransitionTime":"2025-11-27T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.767668 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.767712 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.767723 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.767740 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.767751 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:44Z","lastTransitionTime":"2025-11-27T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.870497 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.870539 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.870547 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.870561 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.870570 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:44Z","lastTransitionTime":"2025-11-27T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.973287 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.973337 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.973358 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.973384 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:44 crc kubenswrapper[4806]: I1127 10:22:44.973404 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:44Z","lastTransitionTime":"2025-11-27T10:22:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.076582 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.076650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.076677 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.076709 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.076775 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:45Z","lastTransitionTime":"2025-11-27T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.116432 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:45 crc kubenswrapper[4806]: E1127 10:22:45.116756 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.127805 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.178603 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.178634 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.178644 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.178658 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.178669 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:45Z","lastTransitionTime":"2025-11-27T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.280925 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.280965 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.280977 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.280998 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.281010 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:45Z","lastTransitionTime":"2025-11-27T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.382982 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.383010 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.383019 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.383033 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.383041 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:45Z","lastTransitionTime":"2025-11-27T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.485099 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.485145 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.485177 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.485198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.485211 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:45Z","lastTransitionTime":"2025-11-27T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.557602 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5cx6g_39930cf4-5f3c-42cc-9384-d366bb1d6a78/kube-multus/0.log" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.557645 4806 generic.go:334] "Generic (PLEG): container finished" podID="39930cf4-5f3c-42cc-9384-d366bb1d6a78" containerID="6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2" exitCode=1 Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.557718 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5cx6g" event={"ID":"39930cf4-5f3c-42cc-9384-d366bb1d6a78","Type":"ContainerDied","Data":"6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2"} Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.558031 4806 scope.go:117] "RemoveContainer" containerID="6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.569356 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.577571 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b7e1203-9780-4d42-8b25-ed6f46baa0af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a7df8a04a47878ce0bd708ebe5559eb0a1e784e3c53749d1790741e8713046c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.587339 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.587369 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.587380 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.587395 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.587405 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:45Z","lastTransitionTime":"2025-11-27T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.593834 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.604586 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.614054 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.629926 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.641899 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:45Z\\\",\\\"message\\\":\\\"2025-11-27T10:22:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298\\\\n2025-11-27T10:22:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298 to /host/opt/cni/bin/\\\\n2025-11-27T10:22:00Z [verbose] multus-daemon started\\\\n2025-11-27T10:22:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T10:22:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.654710 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.673192 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.689399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.689435 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.689446 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.689463 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.689483 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:45Z","lastTransitionTime":"2025-11-27T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.689953 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.701475 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.725024 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:24Z\\\",\\\"message\\\":\\\":[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:24.421527 6328 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"d4efc4a8-c514-4a6b-901c-2953978b50d3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.740055 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.752153 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93396449-9bb1-4cfc-bba4-5c00c92cdf6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6784c0e140b36ee0c0a9135631f867702e99bd54be29abcea1bd7d97f5c533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2d5c2fdd37029ccfa325eb3a2f2fa6d1bcb65b3847dd4997e2cc12798603c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe72cadf6271151c95f0efb2cd1e154c78d1398d9f104c57acc255ecc99738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.768021 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.779697 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.791011 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.792332 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.792376 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.792388 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.792403 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.792435 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:45Z","lastTransitionTime":"2025-11-27T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.807871 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.819124 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:45Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.894678 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.894712 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.894720 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.894733 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.894742 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:45Z","lastTransitionTime":"2025-11-27T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.996868 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.996908 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.996918 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.996934 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:45 crc kubenswrapper[4806]: I1127 10:22:45.996945 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:45Z","lastTransitionTime":"2025-11-27T10:22:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.099471 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.099533 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.099552 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.099575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.099597 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:46Z","lastTransitionTime":"2025-11-27T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.115657 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:46 crc kubenswrapper[4806]: E1127 10:22:46.115800 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.115884 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:46 crc kubenswrapper[4806]: E1127 10:22:46.116055 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.116479 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:46 crc kubenswrapper[4806]: E1127 10:22:46.116606 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.137359 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.153091 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.164462 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.178307 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.187133 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.197346 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93396449-9bb1-4cfc-bba4-5c00c92cdf6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6784c0e140b36ee0c0a9135631f867702e99bd54be29abcea1bd7d97f5c533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2d5c2fdd37029ccfa325eb3a2f2fa6d1bcb65b3847dd4997e2cc12798603c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe72cadf6271151c95f0efb2cd1e154c78d1398d9f104c57acc255ecc99738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.202133 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.202170 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.202180 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.202196 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.202206 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:46Z","lastTransitionTime":"2025-11-27T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.217594 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.237257 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.248024 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.258072 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.268680 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:45Z\\\",\\\"message\\\":\\\"2025-11-27T10:22:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298\\\\n2025-11-27T10:22:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298 to /host/opt/cni/bin/\\\\n2025-11-27T10:22:00Z [verbose] multus-daemon started\\\\n2025-11-27T10:22:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T10:22:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.278632 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.287493 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b7e1203-9780-4d42-8b25-ed6f46baa0af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a7df8a04a47878ce0bd708ebe5559eb0a1e784e3c53749d1790741e8713046c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.299473 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.304393 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.304427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.304439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.304454 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.304467 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:46Z","lastTransitionTime":"2025-11-27T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.313006 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.324596 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.346940 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:24Z\\\",\\\"message\\\":\\\":[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:24.421527 6328 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"d4efc4a8-c514-4a6b-901c-2953978b50d3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.359708 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.372287 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.407145 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.407201 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.407217 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.407260 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.407278 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:46Z","lastTransitionTime":"2025-11-27T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.510005 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.510044 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.510055 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.510071 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.510083 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:46Z","lastTransitionTime":"2025-11-27T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.562590 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5cx6g_39930cf4-5f3c-42cc-9384-d366bb1d6a78/kube-multus/0.log" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.562646 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5cx6g" event={"ID":"39930cf4-5f3c-42cc-9384-d366bb1d6a78","Type":"ContainerStarted","Data":"04e5f8cbf4179151c6868e9168364cc8fa754b47829422ce6a64fa25edf7a949"} Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.577625 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.591562 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e5f8cbf4179151c6868e9168364cc8fa754b47829422ce6a64fa25edf7a949\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:45Z\\\",\\\"message\\\":\\\"2025-11-27T10:22:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298\\\\n2025-11-27T10:22:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298 to /host/opt/cni/bin/\\\\n2025-11-27T10:22:00Z [verbose] multus-daemon started\\\\n2025-11-27T10:22:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T10:22:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.602799 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.615103 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.615149 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.615161 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.615176 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.615188 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:46Z","lastTransitionTime":"2025-11-27T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.620910 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b7e1203-9780-4d42-8b25-ed6f46baa0af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a7df8a04a47878ce0bd708ebe5559eb0a1e784e3c53749d1790741e8713046c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.647862 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.660180 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.671335 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.682831 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.693783 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.704085 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.713934 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.717419 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.717476 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.717489 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.717506 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.717517 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:46Z","lastTransitionTime":"2025-11-27T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.731525 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:24Z\\\",\\\"message\\\":\\\":[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:24.421527 6328 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"d4efc4a8-c514-4a6b-901c-2953978b50d3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.742125 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.755889 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.765916 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.777266 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93396449-9bb1-4cfc-bba4-5c00c92cdf6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6784c0e140b36ee0c0a9135631f867702e99bd54be29abcea1bd7d97f5c533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2d5c2fdd37029ccfa325eb3a2f2fa6d1bcb65b3847dd4997e2cc12798603c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe72cadf6271151c95f0efb2cd1e154c78d1398d9f104c57acc255ecc99738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.788955 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.801418 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.811214 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:46Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.819693 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.819733 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.819745 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.819763 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.819776 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:46Z","lastTransitionTime":"2025-11-27T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.922154 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.922204 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.922217 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.922253 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:46 crc kubenswrapper[4806]: I1127 10:22:46.922267 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:46Z","lastTransitionTime":"2025-11-27T10:22:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.024609 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.024652 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.024674 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.024691 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.024703 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:47Z","lastTransitionTime":"2025-11-27T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.116098 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:47 crc kubenswrapper[4806]: E1127 10:22:47.116261 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.126902 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.126946 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.126959 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.126988 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.127002 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:47Z","lastTransitionTime":"2025-11-27T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.229390 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.229453 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.229467 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.229487 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.229501 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:47Z","lastTransitionTime":"2025-11-27T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.331842 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.331887 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.331925 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.331946 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.331963 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:47Z","lastTransitionTime":"2025-11-27T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.435198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.435273 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.435283 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.435298 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.435309 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:47Z","lastTransitionTime":"2025-11-27T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.537646 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.537697 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.537708 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.537721 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.537730 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:47Z","lastTransitionTime":"2025-11-27T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.640376 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.640422 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.640434 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.640452 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.640463 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:47Z","lastTransitionTime":"2025-11-27T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.742274 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.742306 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.742340 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.742357 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.742370 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:47Z","lastTransitionTime":"2025-11-27T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.844858 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.844882 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.844907 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.844922 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.844931 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:47Z","lastTransitionTime":"2025-11-27T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.946748 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.946788 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.946799 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.946814 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:47 crc kubenswrapper[4806]: I1127 10:22:47.946826 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:47Z","lastTransitionTime":"2025-11-27T10:22:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.049013 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.049051 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.049064 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.049078 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.049089 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:48Z","lastTransitionTime":"2025-11-27T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.115660 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.115677 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:48 crc kubenswrapper[4806]: E1127 10:22:48.115768 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.115879 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:48 crc kubenswrapper[4806]: E1127 10:22:48.115955 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:48 crc kubenswrapper[4806]: E1127 10:22:48.116030 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.151276 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.151318 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.151327 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.151342 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.151354 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:48Z","lastTransitionTime":"2025-11-27T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.253622 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.253650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.253658 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.253670 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.253678 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:48Z","lastTransitionTime":"2025-11-27T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.356586 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.356659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.356669 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.356687 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.356699 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:48Z","lastTransitionTime":"2025-11-27T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.460563 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.460638 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.460660 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.460689 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.460710 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:48Z","lastTransitionTime":"2025-11-27T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.563041 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.563082 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.563091 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.563105 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.563116 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:48Z","lastTransitionTime":"2025-11-27T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.665925 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.665960 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.665969 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.665981 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.665990 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:48Z","lastTransitionTime":"2025-11-27T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.768529 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.768589 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.768600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.768620 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.768630 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:48Z","lastTransitionTime":"2025-11-27T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.870778 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.870840 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.870859 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.870883 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.870899 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:48Z","lastTransitionTime":"2025-11-27T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.973532 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.973582 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.973600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.973624 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:48 crc kubenswrapper[4806]: I1127 10:22:48.973640 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:48Z","lastTransitionTime":"2025-11-27T10:22:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.076018 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.076107 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.076126 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.076153 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.076172 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:49Z","lastTransitionTime":"2025-11-27T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.115615 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:49 crc kubenswrapper[4806]: E1127 10:22:49.115764 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.178630 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.178675 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.178687 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.178704 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.178716 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:49Z","lastTransitionTime":"2025-11-27T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.281177 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.281250 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.281265 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.281282 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.281294 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:49Z","lastTransitionTime":"2025-11-27T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.383949 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.384010 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.384027 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.384049 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.384066 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:49Z","lastTransitionTime":"2025-11-27T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.486411 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.486452 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.486462 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.486477 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.486488 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:49Z","lastTransitionTime":"2025-11-27T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.588640 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.588701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.588719 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.588752 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.588771 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:49Z","lastTransitionTime":"2025-11-27T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.691079 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.691153 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.691171 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.691195 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.691212 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:49Z","lastTransitionTime":"2025-11-27T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.794258 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.794312 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.794335 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.794363 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.794384 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:49Z","lastTransitionTime":"2025-11-27T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.898062 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.898114 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.898136 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.898166 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:49 crc kubenswrapper[4806]: I1127 10:22:49.898186 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:49Z","lastTransitionTime":"2025-11-27T10:22:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.001215 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.001284 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.001297 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.001316 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.001327 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:50Z","lastTransitionTime":"2025-11-27T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.104028 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.104072 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.104080 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.104094 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.104105 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:50Z","lastTransitionTime":"2025-11-27T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.118597 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:50 crc kubenswrapper[4806]: E1127 10:22:50.118693 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.118815 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:50 crc kubenswrapper[4806]: E1127 10:22:50.118859 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.118953 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:50 crc kubenswrapper[4806]: E1127 10:22:50.118998 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.207115 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.207183 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.207205 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.207265 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.207289 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:50Z","lastTransitionTime":"2025-11-27T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.309599 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.309647 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.309664 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.309686 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.309703 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:50Z","lastTransitionTime":"2025-11-27T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.412800 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.412847 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.412873 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.412900 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.412919 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:50Z","lastTransitionTime":"2025-11-27T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.515034 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.515090 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.515106 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.515129 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.515145 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:50Z","lastTransitionTime":"2025-11-27T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.618205 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.618287 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.618304 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.618328 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.618348 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:50Z","lastTransitionTime":"2025-11-27T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.722283 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.722593 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.722776 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.722951 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.723079 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:50Z","lastTransitionTime":"2025-11-27T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.826024 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.826084 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.826101 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.826128 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.826147 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:50Z","lastTransitionTime":"2025-11-27T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.929817 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.929886 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.929904 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.929928 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:50 crc kubenswrapper[4806]: I1127 10:22:50.929946 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:50Z","lastTransitionTime":"2025-11-27T10:22:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.033693 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.033733 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.033744 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.033759 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.033770 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:51Z","lastTransitionTime":"2025-11-27T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.116210 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:51 crc kubenswrapper[4806]: E1127 10:22:51.116383 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.137127 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.137193 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.137209 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.137255 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.137272 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:51Z","lastTransitionTime":"2025-11-27T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.240445 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.240499 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.240513 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.240532 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.240544 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:51Z","lastTransitionTime":"2025-11-27T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.342998 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.343077 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.343128 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.343149 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.343162 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:51Z","lastTransitionTime":"2025-11-27T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.445975 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.446021 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.446064 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.446085 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.446097 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:51Z","lastTransitionTime":"2025-11-27T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.548939 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.549010 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.549027 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.549053 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.549071 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:51Z","lastTransitionTime":"2025-11-27T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.652215 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.652568 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.652668 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.652759 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.652835 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:51Z","lastTransitionTime":"2025-11-27T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.756009 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.756042 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.756056 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.756072 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.756084 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:51Z","lastTransitionTime":"2025-11-27T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.859201 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.859581 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.859750 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.859901 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.860035 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:51Z","lastTransitionTime":"2025-11-27T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.962477 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.962534 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.962545 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.962564 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:51 crc kubenswrapper[4806]: I1127 10:22:51.962575 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:51Z","lastTransitionTime":"2025-11-27T10:22:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.066031 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.066089 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.066097 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.066112 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.066121 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:52Z","lastTransitionTime":"2025-11-27T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.116595 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.116623 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:52 crc kubenswrapper[4806]: E1127 10:22:52.116864 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:52 crc kubenswrapper[4806]: E1127 10:22:52.116747 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.116623 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:52 crc kubenswrapper[4806]: E1127 10:22:52.117512 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.168084 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.168115 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.168124 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.168136 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.168144 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:52Z","lastTransitionTime":"2025-11-27T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.271334 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.271434 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.271458 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.271487 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.271510 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:52Z","lastTransitionTime":"2025-11-27T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.374821 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.374880 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.374900 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.374924 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.374940 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:52Z","lastTransitionTime":"2025-11-27T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.477681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.477750 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.477772 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.477800 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.477822 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:52Z","lastTransitionTime":"2025-11-27T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.581393 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.581682 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.581854 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.582055 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.582286 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:52Z","lastTransitionTime":"2025-11-27T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.685824 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.685884 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.685905 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.685932 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.685953 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:52Z","lastTransitionTime":"2025-11-27T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.788935 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.789342 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.789504 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.789693 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.789859 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:52Z","lastTransitionTime":"2025-11-27T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.893449 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.893508 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.893525 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.893549 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:52 crc kubenswrapper[4806]: I1127 10:22:52.893566 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:52Z","lastTransitionTime":"2025-11-27T10:22:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.013449 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.013514 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.013541 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.013599 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.013625 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.115970 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:53 crc kubenswrapper[4806]: E1127 10:22:53.116180 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.117428 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.117483 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.117501 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.117527 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.117546 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.220606 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.220701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.220725 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.220805 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.220830 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.323058 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.323927 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.324109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.324318 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.324475 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.427808 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.427850 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.427861 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.427880 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.427894 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.432005 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.432051 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.432067 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.432083 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.432095 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: E1127 10:22:53.454481 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:53Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.460453 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.460505 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.460516 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.460535 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.460550 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: E1127 10:22:53.478191 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:53Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.484662 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.484724 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.484742 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.484772 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.484792 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: E1127 10:22:53.507086 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:53Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.513595 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.513650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.513667 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.513692 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.513709 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: E1127 10:22:53.534511 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:53Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.539127 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.539153 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.539163 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.539177 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.539187 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: E1127 10:22:53.559940 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:53Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:53 crc kubenswrapper[4806]: E1127 10:22:53.560075 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.562652 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.562682 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.562691 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.562706 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.562718 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.665130 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.665167 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.665176 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.665190 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.665199 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.767220 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.767262 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.767271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.767284 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.767293 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.869172 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.869216 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.869253 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.869271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.869283 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.971772 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.971842 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.971866 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.971903 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:53 crc kubenswrapper[4806]: I1127 10:22:53.971927 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:53Z","lastTransitionTime":"2025-11-27T10:22:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.075399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.075468 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.075485 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.075510 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.075528 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:54Z","lastTransitionTime":"2025-11-27T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.116346 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.116590 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.116666 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:54 crc kubenswrapper[4806]: E1127 10:22:54.116751 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:54 crc kubenswrapper[4806]: E1127 10:22:54.116588 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:54 crc kubenswrapper[4806]: E1127 10:22:54.116870 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.179076 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.179138 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.179155 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.179181 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.179200 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:54Z","lastTransitionTime":"2025-11-27T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.282932 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.282990 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.283006 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.283028 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.283044 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:54Z","lastTransitionTime":"2025-11-27T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.386364 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.386414 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.386430 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.386455 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.386485 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:54Z","lastTransitionTime":"2025-11-27T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.489964 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.490038 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.490061 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.490084 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.490100 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:54Z","lastTransitionTime":"2025-11-27T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.593017 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.593073 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.593090 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.593112 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.593131 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:54Z","lastTransitionTime":"2025-11-27T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.697709 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.697781 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.697802 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.697826 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.697853 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:54Z","lastTransitionTime":"2025-11-27T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.800427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.800613 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.800634 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.800656 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.800674 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:54Z","lastTransitionTime":"2025-11-27T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.903783 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.903845 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.903863 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.903887 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:54 crc kubenswrapper[4806]: I1127 10:22:54.903904 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:54Z","lastTransitionTime":"2025-11-27T10:22:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.007016 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.007066 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.007080 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.007098 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.007109 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:55Z","lastTransitionTime":"2025-11-27T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.110600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.110664 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.110683 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.110709 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.110727 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:55Z","lastTransitionTime":"2025-11-27T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.116119 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:55 crc kubenswrapper[4806]: E1127 10:22:55.116367 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.213650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.214071 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.214098 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.214120 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.214136 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:55Z","lastTransitionTime":"2025-11-27T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.317334 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.317374 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.317384 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.317398 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.317409 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:55Z","lastTransitionTime":"2025-11-27T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.420310 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.420403 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.420416 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.420436 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.420455 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:55Z","lastTransitionTime":"2025-11-27T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.523140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.523179 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.523190 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.523205 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.523215 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:55Z","lastTransitionTime":"2025-11-27T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.625371 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.625418 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.625426 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.625440 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.625450 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:55Z","lastTransitionTime":"2025-11-27T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.729278 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.729322 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.729330 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.729347 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.729361 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:55Z","lastTransitionTime":"2025-11-27T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.831649 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.831698 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.831710 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.831728 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.831744 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:55Z","lastTransitionTime":"2025-11-27T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.935220 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.935305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.935324 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.935349 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:55 crc kubenswrapper[4806]: I1127 10:22:55.935367 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:55Z","lastTransitionTime":"2025-11-27T10:22:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.037344 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.037388 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.037398 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.037413 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.037423 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:56Z","lastTransitionTime":"2025-11-27T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.116477 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.116562 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.116562 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:56 crc kubenswrapper[4806]: E1127 10:22:56.117217 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:56 crc kubenswrapper[4806]: E1127 10:22:56.117356 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:56 crc kubenswrapper[4806]: E1127 10:22:56.117419 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.117622 4806 scope.go:117] "RemoveContainer" containerID="eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.134391 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.143217 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.143277 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.143292 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.143314 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.143336 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:56Z","lastTransitionTime":"2025-11-27T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.146780 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.162339 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93396449-9bb1-4cfc-bba4-5c00c92cdf6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6784c0e140b36ee0c0a9135631f867702e99bd54be29abcea1bd7d97f5c533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2d5c2fdd37029ccfa325eb3a2f2fa6d1bcb65b3847dd4997e2cc12798603c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe72cadf6271151c95f0efb2cd1e154c78d1398d9f104c57acc255ecc99738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.179496 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.196206 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.210276 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.221511 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.234485 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e5f8cbf4179151c6868e9168364cc8fa754b47829422ce6a64fa25edf7a949\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:45Z\\\",\\\"message\\\":\\\"2025-11-27T10:22:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298\\\\n2025-11-27T10:22:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298 to /host/opt/cni/bin/\\\\n2025-11-27T10:22:00Z [verbose] multus-daemon started\\\\n2025-11-27T10:22:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T10:22:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.243293 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.244711 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.244737 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.244747 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.244764 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.244776 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:56Z","lastTransitionTime":"2025-11-27T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.253360 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b7e1203-9780-4d42-8b25-ed6f46baa0af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a7df8a04a47878ce0bd708ebe5559eb0a1e784e3c53749d1790741e8713046c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.276146 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.293497 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.305502 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.320291 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.335878 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.346603 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.346650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.346658 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.346671 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.346679 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:56Z","lastTransitionTime":"2025-11-27T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.347447 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.355899 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.371595 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:24Z\\\",\\\"message\\\":\\\":[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:24.421527 6328 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"d4efc4a8-c514-4a6b-901c-2953978b50d3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.381264 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.449599 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.449635 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.449646 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.449659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.449670 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:56Z","lastTransitionTime":"2025-11-27T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.553040 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.553096 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.553122 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.553151 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.553173 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:56Z","lastTransitionTime":"2025-11-27T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.607643 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/2.log" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.609935 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerStarted","Data":"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084"} Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.611161 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.631435 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.682965 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.683003 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.683014 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.683031 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.683041 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:56Z","lastTransitionTime":"2025-11-27T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.684031 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.699625 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.711858 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93396449-9bb1-4cfc-bba4-5c00c92cdf6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6784c0e140b36ee0c0a9135631f867702e99bd54be29abcea1bd7d97f5c533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2d5c2fdd37029ccfa325eb3a2f2fa6d1bcb65b3847dd4997e2cc12798603c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe72cadf6271151c95f0efb2cd1e154c78d1398d9f104c57acc255ecc99738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.725343 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.736704 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.746547 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.756271 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.770534 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e5f8cbf4179151c6868e9168364cc8fa754b47829422ce6a64fa25edf7a949\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:45Z\\\",\\\"message\\\":\\\"2025-11-27T10:22:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298\\\\n2025-11-27T10:22:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298 to /host/opt/cni/bin/\\\\n2025-11-27T10:22:00Z [verbose] multus-daemon started\\\\n2025-11-27T10:22:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T10:22:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.780044 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.787456 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b7e1203-9780-4d42-8b25-ed6f46baa0af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a7df8a04a47878ce0bd708ebe5559eb0a1e784e3c53749d1790741e8713046c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.795846 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.795923 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.795942 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.795970 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.796005 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:56Z","lastTransitionTime":"2025-11-27T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.803987 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.816842 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.828661 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.842417 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.854059 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.868144 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.879804 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.897796 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:24Z\\\",\\\"message\\\":\\\":[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:24.421527 6328 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"d4efc4a8-c514-4a6b-901c-2953978b50d3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:56Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.898480 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.898530 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.898545 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.898566 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:56 crc kubenswrapper[4806]: I1127 10:22:56.898580 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:56Z","lastTransitionTime":"2025-11-27T10:22:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.000205 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.000262 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.000274 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.000288 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.000299 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:57Z","lastTransitionTime":"2025-11-27T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.102750 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.102798 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.102814 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.102833 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.102846 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:57Z","lastTransitionTime":"2025-11-27T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.116080 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:57 crc kubenswrapper[4806]: E1127 10:22:57.116220 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.206006 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.206054 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.206069 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.206087 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.206101 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:57Z","lastTransitionTime":"2025-11-27T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.309325 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.309399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.309419 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.309454 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.309479 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:57Z","lastTransitionTime":"2025-11-27T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.411754 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.411800 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.411810 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.411828 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.411845 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:57Z","lastTransitionTime":"2025-11-27T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.514727 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.515223 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.515292 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.515319 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.515337 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:57Z","lastTransitionTime":"2025-11-27T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.617149 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/3.log" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.618001 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.618087 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.618114 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.618148 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.618174 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:57Z","lastTransitionTime":"2025-11-27T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.618370 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/2.log" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.621875 4806 generic.go:334] "Generic (PLEG): container finished" podID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerID="faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084" exitCode=1 Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.621934 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084"} Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.621981 4806 scope.go:117] "RemoveContainer" containerID="eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.623449 4806 scope.go:117] "RemoveContainer" containerID="faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084" Nov 27 10:22:57 crc kubenswrapper[4806]: E1127 10:22:57.623880 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.642695 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.661628 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.678349 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.695320 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.707492 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.726449 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.726707 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.726813 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.726922 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.727061 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:57Z","lastTransitionTime":"2025-11-27T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.753337 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea4308d9c5cb1efc3b9531dc65d6e5cbd01321e8313ea9ff9cd66b97d4c5993\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:24Z\\\",\\\"message\\\":\\\":[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:24.421527 6328 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"d4efc4a8-c514-4a6b-901c-2953978b50d3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-multus/multus-admission-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-multus/multus-admission-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.119\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:57Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:57.119109 6710 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 10:22:57.119122 6710 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 10:22:57.119130 6710 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.788046 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.805460 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.815651 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.825600 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93396449-9bb1-4cfc-bba4-5c00c92cdf6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6784c0e140b36ee0c0a9135631f867702e99bd54be29abcea1bd7d97f5c533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2d5c2fdd37029ccfa325eb3a2f2fa6d1bcb65b3847dd4997e2cc12798603c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe72cadf6271151c95f0efb2cd1e154c78d1398d9f104c57acc255ecc99738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.829199 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.829259 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.829273 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.829290 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.829302 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:57Z","lastTransitionTime":"2025-11-27T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.837637 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.853396 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.864284 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.877357 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.890119 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e5f8cbf4179151c6868e9168364cc8fa754b47829422ce6a64fa25edf7a949\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:45Z\\\",\\\"message\\\":\\\"2025-11-27T10:22:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298\\\\n2025-11-27T10:22:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298 to /host/opt/cni/bin/\\\\n2025-11-27T10:22:00Z [verbose] multus-daemon started\\\\n2025-11-27T10:22:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T10:22:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.900727 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.910986 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b7e1203-9780-4d42-8b25-ed6f46baa0af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a7df8a04a47878ce0bd708ebe5559eb0a1e784e3c53749d1790741e8713046c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.928356 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.931888 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.931989 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.932002 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.932018 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.932030 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:57Z","lastTransitionTime":"2025-11-27T10:22:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:57 crc kubenswrapper[4806]: I1127 10:22:57.944527 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:57Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.034886 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.034984 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.035001 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.035024 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.035049 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:58Z","lastTransitionTime":"2025-11-27T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.116071 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.116130 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:58 crc kubenswrapper[4806]: E1127 10:22:58.116294 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.116333 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:58 crc kubenswrapper[4806]: E1127 10:22:58.116464 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:22:58 crc kubenswrapper[4806]: E1127 10:22:58.116566 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.137556 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.137619 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.137642 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.137669 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.137686 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:58Z","lastTransitionTime":"2025-11-27T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.240295 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.240351 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.240367 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.240396 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.240414 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:58Z","lastTransitionTime":"2025-11-27T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.343217 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.343382 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.343411 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.343441 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.343463 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:58Z","lastTransitionTime":"2025-11-27T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.447352 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.447439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.447469 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.447502 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.447530 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:58Z","lastTransitionTime":"2025-11-27T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.550644 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.550694 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.550717 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.550743 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.550763 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:58Z","lastTransitionTime":"2025-11-27T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.629582 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/3.log" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.635636 4806 scope.go:117] "RemoveContainer" containerID="faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084" Nov 27 10:22:58 crc kubenswrapper[4806]: E1127 10:22:58.635881 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.650745 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.653307 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.653368 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.653391 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.653418 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.653443 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:58Z","lastTransitionTime":"2025-11-27T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.683853 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:57Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:57.119109 6710 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 10:22:57.119122 6710 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 10:22:57.119130 6710 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.702832 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.722586 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.742055 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.757517 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.757601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.757627 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.757657 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.757681 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:58Z","lastTransitionTime":"2025-11-27T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.764676 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.784306 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.810217 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.828082 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.849765 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93396449-9bb1-4cfc-bba4-5c00c92cdf6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6784c0e140b36ee0c0a9135631f867702e99bd54be29abcea1bd7d97f5c533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2d5c2fdd37029ccfa325eb3a2f2fa6d1bcb65b3847dd4997e2cc12798603c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe72cadf6271151c95f0efb2cd1e154c78d1398d9f104c57acc255ecc99738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.860496 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.860551 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.860568 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.860592 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.860609 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:58Z","lastTransitionTime":"2025-11-27T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.882641 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.899084 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.919610 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.941876 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.957095 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e5f8cbf4179151c6868e9168364cc8fa754b47829422ce6a64fa25edf7a949\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:45Z\\\",\\\"message\\\":\\\"2025-11-27T10:22:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298\\\\n2025-11-27T10:22:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298 to /host/opt/cni/bin/\\\\n2025-11-27T10:22:00Z [verbose] multus-daemon started\\\\n2025-11-27T10:22:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T10:22:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.963571 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.963639 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.963665 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.963697 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.963720 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:58Z","lastTransitionTime":"2025-11-27T10:22:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.970897 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:58 crc kubenswrapper[4806]: I1127 10:22:58.985391 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b7e1203-9780-4d42-8b25-ed6f46baa0af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a7df8a04a47878ce0bd708ebe5559eb0a1e784e3c53749d1790741e8713046c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.000796 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:58Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.017794 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:22:59Z is after 2025-08-24T17:21:41Z" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.066441 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.066530 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.066558 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.066592 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.066618 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:59Z","lastTransitionTime":"2025-11-27T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.116389 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.116569 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.169203 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.169554 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.169697 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.169840 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.169976 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:59Z","lastTransitionTime":"2025-11-27T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.273829 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.273932 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.273964 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.273994 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.274017 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:59Z","lastTransitionTime":"2025-11-27T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.376464 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.376542 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.376565 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.376595 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.376617 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:59Z","lastTransitionTime":"2025-11-27T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.480619 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.481034 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.481292 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.481549 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.481753 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:59Z","lastTransitionTime":"2025-11-27T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.585030 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.585107 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.585130 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.585158 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.585180 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:59Z","lastTransitionTime":"2025-11-27T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.688251 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.688519 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.688632 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.688773 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.688864 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:59Z","lastTransitionTime":"2025-11-27T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.791203 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.791500 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.791586 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.791714 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.791809 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:59Z","lastTransitionTime":"2025-11-27T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.816128 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.816475 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.816596 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.816566303 +0000 UTC m=+148.403157107 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.894470 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.894537 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.894555 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.894576 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.894601 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:59Z","lastTransitionTime":"2025-11-27T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.917295 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.917548 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.917607 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.917644 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.917779 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.917807 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.917837 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.917866 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.917879 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.917897 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.917899 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.917873 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.917847219 +0000 UTC m=+148.504438013 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.918019 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.917984233 +0000 UTC m=+148.504575047 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.918071 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.918051974 +0000 UTC m=+148.504642868 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 27 10:22:59 crc kubenswrapper[4806]: E1127 10:22:59.918630 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.918587499 +0000 UTC m=+148.505178313 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.997858 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.997909 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.997919 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.997938 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:22:59 crc kubenswrapper[4806]: I1127 10:22:59.997950 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:22:59Z","lastTransitionTime":"2025-11-27T10:22:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.100650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.100688 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.100701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.100715 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.100725 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:00Z","lastTransitionTime":"2025-11-27T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.115882 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.115954 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:00 crc kubenswrapper[4806]: E1127 10:23:00.116002 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.115954 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:00 crc kubenswrapper[4806]: E1127 10:23:00.116121 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:00 crc kubenswrapper[4806]: E1127 10:23:00.116178 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.203472 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.203510 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.203521 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.203536 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.203546 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:00Z","lastTransitionTime":"2025-11-27T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.306312 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.306388 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.306398 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.306413 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.306424 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:00Z","lastTransitionTime":"2025-11-27T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.409062 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.409095 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.409104 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.409118 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.409127 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:00Z","lastTransitionTime":"2025-11-27T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.511162 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.511195 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.511203 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.511218 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.511257 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:00Z","lastTransitionTime":"2025-11-27T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.614129 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.614166 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.614177 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.614191 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.614202 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:00Z","lastTransitionTime":"2025-11-27T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.716979 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.717208 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.717293 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.717324 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.717349 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:00Z","lastTransitionTime":"2025-11-27T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.820370 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.820433 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.820483 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.820517 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.820538 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:00Z","lastTransitionTime":"2025-11-27T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.924806 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.924909 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.924930 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.924957 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:00 crc kubenswrapper[4806]: I1127 10:23:00.924979 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:00Z","lastTransitionTime":"2025-11-27T10:23:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.028768 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.028838 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.028860 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.028888 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.028912 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:01Z","lastTransitionTime":"2025-11-27T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.116177 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:01 crc kubenswrapper[4806]: E1127 10:23:01.116451 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.131956 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.132035 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.132046 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.132156 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.132169 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:01Z","lastTransitionTime":"2025-11-27T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.235388 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.235445 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.235461 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.235485 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.235500 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:01Z","lastTransitionTime":"2025-11-27T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.338599 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.338666 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.338689 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.338714 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.338735 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:01Z","lastTransitionTime":"2025-11-27T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.441955 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.442028 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.442046 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.442069 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.442086 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:01Z","lastTransitionTime":"2025-11-27T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.545491 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.545541 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.545559 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.545584 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.545605 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:01Z","lastTransitionTime":"2025-11-27T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.648159 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.648218 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.648266 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.648293 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.648309 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:01Z","lastTransitionTime":"2025-11-27T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.751731 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.751818 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.751836 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.751895 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.751916 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:01Z","lastTransitionTime":"2025-11-27T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.854879 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.854972 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.854992 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.855017 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.855039 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:01Z","lastTransitionTime":"2025-11-27T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.958253 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.958610 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.958924 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.959224 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:01 crc kubenswrapper[4806]: I1127 10:23:01.959403 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:01Z","lastTransitionTime":"2025-11-27T10:23:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.063224 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.063307 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.063325 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.063353 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.063370 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:02Z","lastTransitionTime":"2025-11-27T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.116039 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:02 crc kubenswrapper[4806]: E1127 10:23:02.116150 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.116320 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:02 crc kubenswrapper[4806]: E1127 10:23:02.116363 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.116583 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:02 crc kubenswrapper[4806]: E1127 10:23:02.116826 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.165457 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.165831 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.166046 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.166211 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.166419 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:02Z","lastTransitionTime":"2025-11-27T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.269043 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.269455 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.269630 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.269774 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.269951 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:02Z","lastTransitionTime":"2025-11-27T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.373685 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.374117 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.374318 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.374489 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.374635 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:02Z","lastTransitionTime":"2025-11-27T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.478225 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.478320 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.478337 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.478361 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.478379 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:02Z","lastTransitionTime":"2025-11-27T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.581067 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.581399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.581411 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.581428 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.581441 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:02Z","lastTransitionTime":"2025-11-27T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.684950 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.685015 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.685034 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.685059 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.685080 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:02Z","lastTransitionTime":"2025-11-27T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.796857 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.796936 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.796954 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.796975 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.796989 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:02Z","lastTransitionTime":"2025-11-27T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.899625 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.899687 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.899704 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.899726 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:02 crc kubenswrapper[4806]: I1127 10:23:02.899745 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:02Z","lastTransitionTime":"2025-11-27T10:23:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.002575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.002641 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.002661 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.002686 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.002706 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.105980 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.106047 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.106065 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.106091 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.106113 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.116207 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:03 crc kubenswrapper[4806]: E1127 10:23:03.116454 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.209127 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.209180 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.209196 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.209219 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.209272 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.312543 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.312650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.312669 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.312704 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.312729 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.416506 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.416565 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.416576 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.416595 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.416606 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.519945 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.520101 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.520127 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.520156 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.520177 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.610936 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.611014 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.611031 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.611058 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.611077 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: E1127 10:23:03.637005 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.644196 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.644308 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.644331 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.644358 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.644378 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: E1127 10:23:03.666227 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.671577 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.671680 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.671700 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.671772 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.671798 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: E1127 10:23:03.693858 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.700113 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.700182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.700200 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.700225 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.700273 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: E1127 10:23:03.721742 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.726670 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.726926 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.727082 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.727276 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.727477 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: E1127 10:23:03.750164 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:03Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:03 crc kubenswrapper[4806]: E1127 10:23:03.750584 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.753998 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.754213 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.754422 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.754573 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.754718 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.858600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.858960 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.859153 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.859436 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.859663 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.962762 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.962837 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.962861 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.962892 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:03 crc kubenswrapper[4806]: I1127 10:23:03.962913 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:03Z","lastTransitionTime":"2025-11-27T10:23:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.066770 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.066939 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.066969 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.067000 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.067022 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:04Z","lastTransitionTime":"2025-11-27T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.115806 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.115890 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.116190 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:04 crc kubenswrapper[4806]: E1127 10:23:04.116615 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:04 crc kubenswrapper[4806]: E1127 10:23:04.116761 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:04 crc kubenswrapper[4806]: E1127 10:23:04.117087 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.169868 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.170306 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.170489 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.170624 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.170755 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:04Z","lastTransitionTime":"2025-11-27T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.273613 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.273973 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.274402 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.274566 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.274934 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:04Z","lastTransitionTime":"2025-11-27T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.378081 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.378138 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.378161 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.378188 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.378213 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:04Z","lastTransitionTime":"2025-11-27T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.481362 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.481434 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.481451 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.481474 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.481492 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:04Z","lastTransitionTime":"2025-11-27T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.584922 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.584968 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.584986 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.585010 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.585027 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:04Z","lastTransitionTime":"2025-11-27T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.686665 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.686706 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.686716 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.686729 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.686739 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:04Z","lastTransitionTime":"2025-11-27T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.789717 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.789781 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.789802 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.789831 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.789852 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:04Z","lastTransitionTime":"2025-11-27T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.892424 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.892476 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.892496 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.892515 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.892528 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:04Z","lastTransitionTime":"2025-11-27T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.994882 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.994946 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.994958 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.994971 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:04 crc kubenswrapper[4806]: I1127 10:23:04.994983 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:04Z","lastTransitionTime":"2025-11-27T10:23:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.097608 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.097654 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.097665 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.097682 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.097694 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:05Z","lastTransitionTime":"2025-11-27T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.116303 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:05 crc kubenswrapper[4806]: E1127 10:23:05.116406 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.199708 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.199738 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.199754 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.199766 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.199838 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:05Z","lastTransitionTime":"2025-11-27T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.302857 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.302926 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.302944 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.302966 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.302982 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:05Z","lastTransitionTime":"2025-11-27T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.406449 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.406511 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.406534 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.406566 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.406661 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:05Z","lastTransitionTime":"2025-11-27T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.509271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.509301 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.509309 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.509320 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.509328 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:05Z","lastTransitionTime":"2025-11-27T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.610684 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.610755 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.610778 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.610803 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.610825 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:05Z","lastTransitionTime":"2025-11-27T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.713079 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.713134 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.713149 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.713171 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.713186 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:05Z","lastTransitionTime":"2025-11-27T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.815849 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.815954 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.815968 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.815985 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.815996 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:05Z","lastTransitionTime":"2025-11-27T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.918597 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.918663 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.918685 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.918711 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:05 crc kubenswrapper[4806]: I1127 10:23:05.918731 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:05Z","lastTransitionTime":"2025-11-27T10:23:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.021447 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.021507 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.021525 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.021550 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.021571 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:06Z","lastTransitionTime":"2025-11-27T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.116167 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.116294 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:06 crc kubenswrapper[4806]: E1127 10:23:06.116307 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:06 crc kubenswrapper[4806]: E1127 10:23:06.116412 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.116651 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:06 crc kubenswrapper[4806]: E1127 10:23:06.116754 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.123264 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.123292 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.123300 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.123313 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.123321 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:06Z","lastTransitionTime":"2025-11-27T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.135604 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.152751 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.175221 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.188977 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.212454 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:57Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:57.119109 6710 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 10:22:57.119122 6710 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 10:22:57.119130 6710 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.225950 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.226041 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.226068 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.226098 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.226120 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:06Z","lastTransitionTime":"2025-11-27T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.232586 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.249529 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.263017 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93396449-9bb1-4cfc-bba4-5c00c92cdf6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6784c0e140b36ee0c0a9135631f867702e99bd54be29abcea1bd7d97f5c533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2d5c2fdd37029ccfa325eb3a2f2fa6d1bcb65b3847dd4997e2cc12798603c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe72cadf6271151c95f0efb2cd1e154c78d1398d9f104c57acc255ecc99738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.278928 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.298029 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.315814 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.329575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.329632 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.329651 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.329675 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.329692 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:06Z","lastTransitionTime":"2025-11-27T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.340501 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.355568 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e5f8cbf4179151c6868e9168364cc8fa754b47829422ce6a64fa25edf7a949\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:45Z\\\",\\\"message\\\":\\\"2025-11-27T10:22:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298\\\\n2025-11-27T10:22:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298 to /host/opt/cni/bin/\\\\n2025-11-27T10:22:00Z [verbose] multus-daemon started\\\\n2025-11-27T10:22:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T10:22:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.364849 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.372683 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b7e1203-9780-4d42-8b25-ed6f46baa0af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a7df8a04a47878ce0bd708ebe5559eb0a1e784e3c53749d1790741e8713046c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.395478 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.405922 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.418328 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.428892 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:06Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.431456 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.431483 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.431490 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.431502 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.431511 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:06Z","lastTransitionTime":"2025-11-27T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.534940 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.535252 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.535352 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.535440 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.535542 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:06Z","lastTransitionTime":"2025-11-27T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.637670 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.638080 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.638142 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.638208 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.638320 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:06Z","lastTransitionTime":"2025-11-27T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.742563 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.742641 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.742659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.742691 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.742710 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:06Z","lastTransitionTime":"2025-11-27T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.846691 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.846756 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.846774 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.846802 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.846822 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:06Z","lastTransitionTime":"2025-11-27T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.949502 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.949552 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.949562 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.949574 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:06 crc kubenswrapper[4806]: I1127 10:23:06.949583 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:06Z","lastTransitionTime":"2025-11-27T10:23:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.052852 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.052934 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.052948 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.052978 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.052997 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:07Z","lastTransitionTime":"2025-11-27T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.116125 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:07 crc kubenswrapper[4806]: E1127 10:23:07.116386 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.156097 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.156152 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.156172 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.156196 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.156214 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:07Z","lastTransitionTime":"2025-11-27T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.259678 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.259743 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.259755 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.259777 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.259795 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:07Z","lastTransitionTime":"2025-11-27T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.363008 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.363118 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.363140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.363171 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.363197 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:07Z","lastTransitionTime":"2025-11-27T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.466593 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.466659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.466671 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.466685 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.466695 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:07Z","lastTransitionTime":"2025-11-27T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.569895 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.569945 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.569963 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.569999 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.570034 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:07Z","lastTransitionTime":"2025-11-27T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.673728 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.673795 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.673816 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.673846 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.673871 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:07Z","lastTransitionTime":"2025-11-27T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.778014 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.778068 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.778085 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.778107 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.778124 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:07Z","lastTransitionTime":"2025-11-27T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.882092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.882176 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.882202 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.882271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.882342 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:07Z","lastTransitionTime":"2025-11-27T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.984911 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.984957 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.984969 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.984986 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:07 crc kubenswrapper[4806]: I1127 10:23:07.985000 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:07Z","lastTransitionTime":"2025-11-27T10:23:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.088200 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.088306 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.088339 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.088382 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.088406 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:08Z","lastTransitionTime":"2025-11-27T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.116417 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.116451 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.116556 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:08 crc kubenswrapper[4806]: E1127 10:23:08.116670 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:08 crc kubenswrapper[4806]: E1127 10:23:08.116864 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:08 crc kubenswrapper[4806]: E1127 10:23:08.116950 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.191820 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.191875 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.191892 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.191914 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.191931 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:08Z","lastTransitionTime":"2025-11-27T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.294503 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.294598 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.294611 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.294628 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.294639 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:08Z","lastTransitionTime":"2025-11-27T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.398141 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.398217 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.398291 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.398331 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.398354 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:08Z","lastTransitionTime":"2025-11-27T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.501904 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.501966 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.501990 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.502020 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.502041 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:08Z","lastTransitionTime":"2025-11-27T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.606864 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.606954 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.607020 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.607054 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.607070 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:08Z","lastTransitionTime":"2025-11-27T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.711101 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.711147 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.711158 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.711176 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.711188 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:08Z","lastTransitionTime":"2025-11-27T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.814604 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.814672 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.814690 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.814716 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.814736 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:08Z","lastTransitionTime":"2025-11-27T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.917614 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.917651 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.917677 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.917695 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:08 crc kubenswrapper[4806]: I1127 10:23:08.917705 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:08Z","lastTransitionTime":"2025-11-27T10:23:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.020439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.020508 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.020529 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.020561 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.020587 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:09Z","lastTransitionTime":"2025-11-27T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.116641 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:09 crc kubenswrapper[4806]: E1127 10:23:09.117424 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.124120 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.124177 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.124189 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.124211 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.124251 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:09Z","lastTransitionTime":"2025-11-27T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.227942 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.228003 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.228019 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.228048 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.228064 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:09Z","lastTransitionTime":"2025-11-27T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.331436 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.331469 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.331478 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.331491 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.331500 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:09Z","lastTransitionTime":"2025-11-27T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.433973 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.434030 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.434047 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.434071 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.434087 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:09Z","lastTransitionTime":"2025-11-27T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.536484 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.536552 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.536578 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.536607 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.536628 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:09Z","lastTransitionTime":"2025-11-27T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.639792 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.639859 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.639883 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.639912 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.639932 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:09Z","lastTransitionTime":"2025-11-27T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.742137 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.742197 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.742214 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.742286 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.742314 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:09Z","lastTransitionTime":"2025-11-27T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.845287 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.845357 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.845374 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.845399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.845420 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:09Z","lastTransitionTime":"2025-11-27T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.947734 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.947781 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.947793 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.947814 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:09 crc kubenswrapper[4806]: I1127 10:23:09.947826 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:09Z","lastTransitionTime":"2025-11-27T10:23:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.051835 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.051899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.051917 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.051940 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.051957 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:10Z","lastTransitionTime":"2025-11-27T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.116811 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.116875 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.116927 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:10 crc kubenswrapper[4806]: E1127 10:23:10.117020 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:10 crc kubenswrapper[4806]: E1127 10:23:10.117115 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:10 crc kubenswrapper[4806]: E1127 10:23:10.117349 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.155664 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.155742 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.155760 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.156277 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.156348 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:10Z","lastTransitionTime":"2025-11-27T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.258953 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.259023 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.259042 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.259069 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.259089 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:10Z","lastTransitionTime":"2025-11-27T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.362510 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.362589 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.362615 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.362657 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.362681 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:10Z","lastTransitionTime":"2025-11-27T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.465941 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.466012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.466037 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.466065 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.466084 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:10Z","lastTransitionTime":"2025-11-27T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.569725 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.569769 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.569783 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.569801 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.569815 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:10Z","lastTransitionTime":"2025-11-27T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.672976 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.673051 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.673071 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.673097 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.673116 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:10Z","lastTransitionTime":"2025-11-27T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.776951 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.777038 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.777064 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.777097 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.777119 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:10Z","lastTransitionTime":"2025-11-27T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.880692 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.880766 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.880784 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.880811 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.880835 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:10Z","lastTransitionTime":"2025-11-27T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.984080 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.984149 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.984166 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.984192 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:10 crc kubenswrapper[4806]: I1127 10:23:10.984214 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:10Z","lastTransitionTime":"2025-11-27T10:23:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.088141 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.088185 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.088196 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.088265 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.088281 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:11Z","lastTransitionTime":"2025-11-27T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.116547 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:11 crc kubenswrapper[4806]: E1127 10:23:11.116742 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.191027 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.191053 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.191061 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.191073 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.191081 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:11Z","lastTransitionTime":"2025-11-27T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.293854 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.293922 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.293944 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.293972 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.293993 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:11Z","lastTransitionTime":"2025-11-27T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.396691 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.396722 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.396733 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.396749 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.396761 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:11Z","lastTransitionTime":"2025-11-27T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.499345 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.499409 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.499427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.499452 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.499473 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:11Z","lastTransitionTime":"2025-11-27T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.602630 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.602671 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.602685 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.602701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.602713 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:11Z","lastTransitionTime":"2025-11-27T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.706087 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.706171 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.706182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.706198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.706210 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:11Z","lastTransitionTime":"2025-11-27T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.809191 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.809217 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.809226 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.809253 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.809264 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:11Z","lastTransitionTime":"2025-11-27T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.911647 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.911703 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.911720 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.911742 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:11 crc kubenswrapper[4806]: I1127 10:23:11.911759 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:11Z","lastTransitionTime":"2025-11-27T10:23:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.014250 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.014277 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.014287 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.014306 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.014323 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:12Z","lastTransitionTime":"2025-11-27T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.115699 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.115737 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.115705 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:12 crc kubenswrapper[4806]: E1127 10:23:12.115921 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:12 crc kubenswrapper[4806]: E1127 10:23:12.115968 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:12 crc kubenswrapper[4806]: E1127 10:23:12.116103 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.118144 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.118180 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.118189 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.118205 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.118217 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:12Z","lastTransitionTime":"2025-11-27T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.221151 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.221193 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.221201 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.221216 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.221225 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:12Z","lastTransitionTime":"2025-11-27T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.324687 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.324755 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.324772 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.324797 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.324814 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:12Z","lastTransitionTime":"2025-11-27T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.428178 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.428305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.428334 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.428399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.428423 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:12Z","lastTransitionTime":"2025-11-27T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.531069 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.531131 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.531150 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.531175 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.531192 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:12Z","lastTransitionTime":"2025-11-27T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.634603 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.634665 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.634682 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.634705 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.634723 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:12Z","lastTransitionTime":"2025-11-27T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.738097 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.738173 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.738195 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.738224 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.738293 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:12Z","lastTransitionTime":"2025-11-27T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.841110 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.841183 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.841206 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.841279 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.841313 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:12Z","lastTransitionTime":"2025-11-27T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.944260 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.944331 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.944353 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.944382 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:12 crc kubenswrapper[4806]: I1127 10:23:12.944406 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:12Z","lastTransitionTime":"2025-11-27T10:23:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.047588 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.047658 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.047680 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.047707 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.047729 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:13Z","lastTransitionTime":"2025-11-27T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.116731 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:13 crc kubenswrapper[4806]: E1127 10:23:13.117435 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.117926 4806 scope.go:117] "RemoveContainer" containerID="faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084" Nov 27 10:23:13 crc kubenswrapper[4806]: E1127 10:23:13.118174 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.151675 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.151736 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.151758 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.151787 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.151808 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:13Z","lastTransitionTime":"2025-11-27T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.254667 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.254736 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.254761 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.254789 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.254810 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:13Z","lastTransitionTime":"2025-11-27T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.357209 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.357482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.357493 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.357508 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.357520 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:13Z","lastTransitionTime":"2025-11-27T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.460118 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.460154 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.460166 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.460179 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.460189 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:13Z","lastTransitionTime":"2025-11-27T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.562805 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.562906 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.562932 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.563016 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.563083 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:13Z","lastTransitionTime":"2025-11-27T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.666113 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.666162 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.666174 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.666190 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.666203 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:13Z","lastTransitionTime":"2025-11-27T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.769208 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.769325 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.769345 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.769370 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.769390 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:13Z","lastTransitionTime":"2025-11-27T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.872373 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.872463 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.872479 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.872504 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.872523 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:13Z","lastTransitionTime":"2025-11-27T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.972702 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.973067 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.973309 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.973489 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:13 crc kubenswrapper[4806]: I1127 10:23:13.973630 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:13Z","lastTransitionTime":"2025-11-27T10:23:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:13 crc kubenswrapper[4806]: E1127 10:23:13.994827 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:13Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.001167 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.001265 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.001292 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.001324 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.001348 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:14Z","lastTransitionTime":"2025-11-27T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:14 crc kubenswrapper[4806]: E1127 10:23:14.021545 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:14Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.027102 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.027265 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.027298 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.027328 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.027350 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:14Z","lastTransitionTime":"2025-11-27T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:14 crc kubenswrapper[4806]: E1127 10:23:14.048025 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:14Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.053418 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.053482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.053499 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.053523 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.053631 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:14Z","lastTransitionTime":"2025-11-27T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:14 crc kubenswrapper[4806]: E1127 10:23:14.076220 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:14Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.081751 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.081812 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.081828 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.081851 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.081868 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:14Z","lastTransitionTime":"2025-11-27T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:14 crc kubenswrapper[4806]: E1127 10:23:14.102531 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:14Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:14 crc kubenswrapper[4806]: E1127 10:23:14.102818 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.105283 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.105337 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.105358 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.105383 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.105399 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:14Z","lastTransitionTime":"2025-11-27T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.116635 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.116693 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.116751 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:14 crc kubenswrapper[4806]: E1127 10:23:14.116809 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:14 crc kubenswrapper[4806]: E1127 10:23:14.117005 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:14 crc kubenswrapper[4806]: E1127 10:23:14.117067 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.208096 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.208160 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.208177 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.208200 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.208220 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:14Z","lastTransitionTime":"2025-11-27T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.310853 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.310899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.310912 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.310951 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.310974 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:14Z","lastTransitionTime":"2025-11-27T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.414572 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.414632 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.414648 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.414671 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.414688 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:14Z","lastTransitionTime":"2025-11-27T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.518305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.518373 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.518396 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.518420 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.518440 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:14Z","lastTransitionTime":"2025-11-27T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.632703 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.632771 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.632793 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.632821 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.632841 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:14Z","lastTransitionTime":"2025-11-27T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.736135 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.736182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.736199 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.736223 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.736277 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:14Z","lastTransitionTime":"2025-11-27T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.839349 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.839434 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.839460 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.839493 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.839517 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:14Z","lastTransitionTime":"2025-11-27T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.942457 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.942549 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.942574 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.942899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:14 crc kubenswrapper[4806]: I1127 10:23:14.942918 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:14Z","lastTransitionTime":"2025-11-27T10:23:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.045749 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.045832 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.045854 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.045882 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.045903 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:15Z","lastTransitionTime":"2025-11-27T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.115939 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:15 crc kubenswrapper[4806]: E1127 10:23:15.116145 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.149012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.149921 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.150115 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.150382 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.150539 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:15Z","lastTransitionTime":"2025-11-27T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.253287 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.253353 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.253375 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.253401 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.253420 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:15Z","lastTransitionTime":"2025-11-27T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.356706 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.356771 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.356792 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.356819 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.356841 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:15Z","lastTransitionTime":"2025-11-27T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.459721 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.459786 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.459822 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.459859 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.459883 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:15Z","lastTransitionTime":"2025-11-27T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.561964 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.562029 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.562050 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.562076 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.562094 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:15Z","lastTransitionTime":"2025-11-27T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.665224 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.665307 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.665324 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.665348 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.665366 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:15Z","lastTransitionTime":"2025-11-27T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.768262 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.768537 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.768657 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.768767 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.768855 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:15Z","lastTransitionTime":"2025-11-27T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.801300 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:15 crc kubenswrapper[4806]: E1127 10:23:15.801514 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:23:15 crc kubenswrapper[4806]: E1127 10:23:15.801653 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs podName:3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e nodeName:}" failed. No retries permitted until 2025-11-27 10:24:19.801617072 +0000 UTC m=+164.388207876 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs") pod "network-metrics-daemon-p2gpw" (UID: "3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.872676 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.872719 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.872730 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.872744 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.872753 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:15Z","lastTransitionTime":"2025-11-27T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.976165 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.976217 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.976244 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.976260 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:15 crc kubenswrapper[4806]: I1127 10:23:15.976271 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:15Z","lastTransitionTime":"2025-11-27T10:23:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.079358 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.079419 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.079439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.079475 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.079512 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:16Z","lastTransitionTime":"2025-11-27T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.116505 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.116523 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:16 crc kubenswrapper[4806]: E1127 10:23:16.116946 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.117284 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:16 crc kubenswrapper[4806]: E1127 10:23:16.117530 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:16 crc kubenswrapper[4806]: E1127 10:23:16.117611 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.131812 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a074e78-8e20-4546-892d-8fee6cbf4aa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://738cf55259b1cbbc21d13c5a9e1df779b500f86fa03432e3d58963742cb91f01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2eb16df2ea2bd9faf7ac32345734e9336a4675555a6228372ce3499c7e255a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8bd7c8193f83329d563e73e65c74b837db8f8495bfb2605cfc1b41291f3a2a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.146720 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-68rwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3dfed96b-3244-4631-b080-cf112ad196dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb801769521f8fdebc477d626e6fdb5c753952a33dc462f18622baccc1463311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c9fld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:57Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-68rwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.167673 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8ed6bbc-965a-412c-a3ee-4763035aa34e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:57Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1127 10:22:57.119109 6710 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 10:22:57.119122 6710 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1127 10:22:57.119130 6710 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czmdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rsgmm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.180729 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5d1ac6b-daa1-4c04-a3b3-51ce910ab441\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cda5bee521d7e8fc4c86c0c955678a5cdb1a2a496a10a7285b1a2dab4d7fc21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81583577448a4f76dd94eb8c43b85a9b06b03b1e2e204eb5e14b8607c3ebba34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-g2jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lqb5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.181898 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.182120 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.182638 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.182928 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.183130 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:16Z","lastTransitionTime":"2025-11-27T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.198178 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-44ndp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92d21faf-b0b5-469f-8c6f-af3b0ee6240f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b25f9f4ac0673d3a8a4f788694a734cd75247352c5e10e1d4175d3f2dce00bac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://137e8bd4988c1dc285c53c884d7707a40a787821f99ba79a333f2279f9dd72e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://24fd833b682b7cf046517c6904b975047c0c2abacdefefaba17a24f76763b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd523871e3c2495015e96261654ee67a6ffb69be988693f100756c1cf62a7c0c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca645c2b8e8236244c4b5255ad4a49fdd51520f431fe102e418ae781e658cd9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://011f9ddb07ab80eae0bbed9e9a647f2483b18ac6d38a4561c7871f15dbac22a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f59a5745831b6dfdc695a0d1b3edd38403d47f85b159e42b603ea1b6a1ec59d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:22:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-65952\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-44ndp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.213290 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dzlc2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9f8291c9-c70f-4d08-bb29-fbc523d38fa3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c05fd2c79e834b7385aab6dc10647a10b3ec3f0e7e5cf7e604d13dda7f93fb20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hfw9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dzlc2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.227101 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93396449-9bb1-4cfc-bba4-5c00c92cdf6a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6784c0e140b36ee0c0a9135631f867702e99bd54be29abcea1bd7d97f5c533\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b2d5c2fdd37029ccfa325eb3a2f2fa6d1bcb65b3847dd4997e2cc12798603c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74fe72cadf6271151c95f0efb2cd1e154c78d1398d9f104c57acc255ecc99738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d402082f82afe91bec675add7abf1fde6939c0d9a488dbc2b6af87788a2bf71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.243007 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.260600 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8d02751090a1f37915ed3c7f60b84632ccc6ccdc4d571b6153086013bb78ba2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.270612 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ea0cad1-8e8b-458c-a943-7191b36e35a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://523be30c415daa546c5fdd2b1045578e2770704390c587e2ec630742f7c309a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cknt9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mq9p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.285608 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c8007974f90508fb627f1e1e19c2d8805f4a7b66f46bde2ea43616a4c4aac53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5967f3c5550e2e23e8d9b8dadbd964ea05c44b8c7e9a8e07d22040de0408b539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.286649 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.286672 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.286680 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.286692 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.286701 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:16Z","lastTransitionTime":"2025-11-27T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.298166 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5cx6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39930cf4-5f3c-42cc-9384-d366bb1d6a78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e5f8cbf4179151c6868e9168364cc8fa754b47829422ce6a64fa25edf7a949\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-27T10:22:45Z\\\",\\\"message\\\":\\\"2025-11-27T10:22:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298\\\\n2025-11-27T10:22:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b48cfa88-288b-4834-b491-369c79685298 to /host/opt/cni/bin/\\\\n2025-11-27T10:22:00Z [verbose] multus-daemon started\\\\n2025-11-27T10:22:00Z [verbose] Readiness Indicator file check\\\\n2025-11-27T10:22:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:22:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njxg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:58Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5cx6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.310399 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrmz9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:22:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p2gpw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.324883 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b7e1203-9780-4d42-8b25-ed6f46baa0af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a7df8a04a47878ce0bd708ebe5559eb0a1e784e3c53749d1790741e8713046c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5cb0e18e848e726957fdd809fe88e2bba5423813bcf2f7e30157972574519a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.353975 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b7a07b8-74db-4292-a289-72ebc12337ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37813f470e989359ee5c479ed0df68fd1c2754d41afa49912c65fb559935d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://922a3318cf4aa71f6f18025ec3c0c29cb89cd2ceff7d6970fcf642135317e4b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd92253f08ff2f62cf94d1c86c03b5b1e3551040570729884ee2a36e820797ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fa34e9e4e2965c9230f236c73a89d2aea69eb036aa60fd685b19c15030198de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83190e416d8661f0d396b484d84448a5ddb891b4c43b0bd14ea86d7d4254c707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7376ac5a31a03f812e339f51c938dbc7f9e22e66372c7814c35c687385a515ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c3055fbcc4331bfb1f2f506c74fc5ea992294a4e76c65b69263e179c04e4e70\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b150c59984215d48c65d41a49a30b063f2cb408ef0dddadb4075c76f9997eee4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.370496 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b916085c66e43f182dd0531a0b367bbf6ae81fcd36273d4b56ac708499b8a27b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.388103 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.390130 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.390220 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.390289 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.390389 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.390411 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:16Z","lastTransitionTime":"2025-11-27T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.409801 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:22:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-27T10:21:50Z\\\",\\\"message\\\":\\\"W1127 10:21:39.355388 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1127 10:21:39.355735 1 crypto.go:601] Generating new CA for check-endpoints-signer@1764238899 cert, and key in /tmp/serving-cert-4203248494/serving-signer.crt, /tmp/serving-cert-4203248494/serving-signer.key\\\\nI1127 10:21:39.623778 1 observer_polling.go:159] Starting file observer\\\\nW1127 10:21:39.629765 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1127 10:21:39.629959 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1127 10:21:39.632339 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4203248494/tls.crt::/tmp/serving-cert-4203248494/tls.key\\\\\\\"\\\\nF1127 10:21:50.072123 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-27T10:21:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-27T10:21:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-27T10:21:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-27T10:21:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.425557 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-27T10:21:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:16Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.493681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.493734 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.493754 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.493779 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.493805 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:16Z","lastTransitionTime":"2025-11-27T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.598316 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.598382 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.598403 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.598431 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.598456 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:16Z","lastTransitionTime":"2025-11-27T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.706987 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.707478 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.707672 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.707905 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.708081 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:16Z","lastTransitionTime":"2025-11-27T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.811436 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.811507 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.811530 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.811558 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.811583 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:16Z","lastTransitionTime":"2025-11-27T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.914841 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.914922 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.914955 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.914988 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:16 crc kubenswrapper[4806]: I1127 10:23:16.915013 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:16Z","lastTransitionTime":"2025-11-27T10:23:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.019804 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.019864 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.019887 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.019914 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.019936 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:17Z","lastTransitionTime":"2025-11-27T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.115791 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:17 crc kubenswrapper[4806]: E1127 10:23:17.115964 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.122506 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.122559 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.122567 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.122581 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.122590 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:17Z","lastTransitionTime":"2025-11-27T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.224824 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.224890 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.224898 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.224928 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.224941 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:17Z","lastTransitionTime":"2025-11-27T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.327725 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.327791 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.327803 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.327818 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.327852 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:17Z","lastTransitionTime":"2025-11-27T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.430975 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.431007 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.431018 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.431032 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.431043 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:17Z","lastTransitionTime":"2025-11-27T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.533344 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.533428 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.533454 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.533488 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.533508 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:17Z","lastTransitionTime":"2025-11-27T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.636944 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.637012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.637028 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.637052 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.637069 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:17Z","lastTransitionTime":"2025-11-27T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.741438 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.741496 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.741514 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.741538 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.741555 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:17Z","lastTransitionTime":"2025-11-27T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.844556 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.844629 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.844644 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.844670 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.844693 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:17Z","lastTransitionTime":"2025-11-27T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.948118 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.948189 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.948206 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.948278 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:17 crc kubenswrapper[4806]: I1127 10:23:17.948306 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:17Z","lastTransitionTime":"2025-11-27T10:23:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.051364 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.051515 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.051537 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.051560 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.051578 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:18Z","lastTransitionTime":"2025-11-27T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.116501 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.116600 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:18 crc kubenswrapper[4806]: E1127 10:23:18.116678 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.116759 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:18 crc kubenswrapper[4806]: E1127 10:23:18.116858 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:18 crc kubenswrapper[4806]: E1127 10:23:18.117026 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.154586 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.154636 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.154647 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.154664 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.154681 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:18Z","lastTransitionTime":"2025-11-27T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.257715 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.257762 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.257772 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.257788 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.257799 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:18Z","lastTransitionTime":"2025-11-27T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.360298 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.360338 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.360347 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.360360 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.360368 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:18Z","lastTransitionTime":"2025-11-27T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.464628 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.464668 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.464679 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.464695 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.464706 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:18Z","lastTransitionTime":"2025-11-27T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.567335 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.567381 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.567395 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.567411 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.567423 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:18Z","lastTransitionTime":"2025-11-27T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.670570 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.670643 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.670666 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.670694 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.670716 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:18Z","lastTransitionTime":"2025-11-27T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.773818 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.773868 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.773881 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.773900 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.773917 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:18Z","lastTransitionTime":"2025-11-27T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.876956 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.876992 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.877000 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.877012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.877021 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:18Z","lastTransitionTime":"2025-11-27T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.979872 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.979915 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.979928 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.979943 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:18 crc kubenswrapper[4806]: I1127 10:23:18.979954 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:18Z","lastTransitionTime":"2025-11-27T10:23:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.082901 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.082935 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.082951 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.082967 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.082980 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:19Z","lastTransitionTime":"2025-11-27T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.116464 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:19 crc kubenswrapper[4806]: E1127 10:23:19.116607 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.185352 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.185423 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.185445 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.185475 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.185499 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:19Z","lastTransitionTime":"2025-11-27T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.288784 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.288867 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.288880 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.288898 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.288910 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:19Z","lastTransitionTime":"2025-11-27T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.392109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.392172 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.392191 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.392211 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.392226 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:19Z","lastTransitionTime":"2025-11-27T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.495928 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.496049 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.496069 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.496093 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.496110 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:19Z","lastTransitionTime":"2025-11-27T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.598593 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.598636 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.598646 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.598661 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.598672 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:19Z","lastTransitionTime":"2025-11-27T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.700943 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.700991 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.701007 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.701040 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.701055 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:19Z","lastTransitionTime":"2025-11-27T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.803826 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.803875 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.803893 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.803918 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.803941 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:19Z","lastTransitionTime":"2025-11-27T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.906182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.906301 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.906328 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.906356 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:19 crc kubenswrapper[4806]: I1127 10:23:19.906376 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:19Z","lastTransitionTime":"2025-11-27T10:23:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.009399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.009453 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.009470 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.009492 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.009508 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:20Z","lastTransitionTime":"2025-11-27T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.112306 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.112374 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.112393 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.112444 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.112562 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:20Z","lastTransitionTime":"2025-11-27T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.116481 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.116481 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.116600 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:20 crc kubenswrapper[4806]: E1127 10:23:20.116724 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:20 crc kubenswrapper[4806]: E1127 10:23:20.116869 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:20 crc kubenswrapper[4806]: E1127 10:23:20.117001 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.215287 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.215347 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.215369 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.215398 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.215422 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:20Z","lastTransitionTime":"2025-11-27T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.318748 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.318811 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.318827 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.318853 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.318876 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:20Z","lastTransitionTime":"2025-11-27T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.421371 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.421433 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.421453 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.421476 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.421492 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:20Z","lastTransitionTime":"2025-11-27T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.524496 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.524538 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.524551 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.524567 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.524578 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:20Z","lastTransitionTime":"2025-11-27T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.627637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.627710 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.627728 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.627752 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.627782 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:20Z","lastTransitionTime":"2025-11-27T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.732329 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.732424 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.732448 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.732479 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.732500 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:20Z","lastTransitionTime":"2025-11-27T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.835343 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.835389 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.835405 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.835430 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.835447 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:20Z","lastTransitionTime":"2025-11-27T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.937808 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.937854 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.937871 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.937890 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:20 crc kubenswrapper[4806]: I1127 10:23:20.937905 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:20Z","lastTransitionTime":"2025-11-27T10:23:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.040849 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.041001 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.041021 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.041052 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.041117 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:21Z","lastTransitionTime":"2025-11-27T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.116310 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:21 crc kubenswrapper[4806]: E1127 10:23:21.116555 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.143643 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.143690 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.143705 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.143731 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.143748 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:21Z","lastTransitionTime":"2025-11-27T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.246651 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.246687 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.246703 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.246723 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.246740 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:21Z","lastTransitionTime":"2025-11-27T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.349885 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.349929 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.349943 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.349961 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.349972 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:21Z","lastTransitionTime":"2025-11-27T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.453198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.453274 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.453292 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.453330 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.453350 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:21Z","lastTransitionTime":"2025-11-27T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.556355 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.556419 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.556438 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.556460 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.556478 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:21Z","lastTransitionTime":"2025-11-27T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.659011 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.659045 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.659056 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.659072 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.659083 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:21Z","lastTransitionTime":"2025-11-27T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.761535 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.761588 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.761604 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.761626 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.761642 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:21Z","lastTransitionTime":"2025-11-27T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.864260 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.864306 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.864322 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.864343 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.864359 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:21Z","lastTransitionTime":"2025-11-27T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.968217 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.968338 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.968363 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.968399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:21 crc kubenswrapper[4806]: I1127 10:23:21.968420 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:21Z","lastTransitionTime":"2025-11-27T10:23:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.071823 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.071879 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.071896 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.071919 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.071934 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:22Z","lastTransitionTime":"2025-11-27T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.116331 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.116417 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:22 crc kubenswrapper[4806]: E1127 10:23:22.116481 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.116331 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:22 crc kubenswrapper[4806]: E1127 10:23:22.117501 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:22 crc kubenswrapper[4806]: E1127 10:23:22.117377 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.175759 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.175832 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.175856 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.175886 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.175910 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:22Z","lastTransitionTime":"2025-11-27T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.279025 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.279101 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.279115 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.279137 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.279149 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:22Z","lastTransitionTime":"2025-11-27T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.382389 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.382456 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.382475 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.382503 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.382519 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:22Z","lastTransitionTime":"2025-11-27T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.486663 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.486718 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.486728 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.486747 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.486761 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:22Z","lastTransitionTime":"2025-11-27T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.589913 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.590356 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.590384 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.590411 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.590429 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:22Z","lastTransitionTime":"2025-11-27T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.693575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.693653 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.693675 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.693720 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.693749 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:22Z","lastTransitionTime":"2025-11-27T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.796292 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.796346 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.796362 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.796385 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.796400 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:22Z","lastTransitionTime":"2025-11-27T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.899654 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.899704 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.899719 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.899744 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:22 crc kubenswrapper[4806]: I1127 10:23:22.899759 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:22Z","lastTransitionTime":"2025-11-27T10:23:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:23 crc kubenswrapper[4806]: I1127 10:23:23.001998 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:23 crc kubenswrapper[4806]: I1127 10:23:23.002038 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:23 crc kubenswrapper[4806]: I1127 10:23:23.002049 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:23 crc kubenswrapper[4806]: I1127 10:23:23.002064 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:23 crc kubenswrapper[4806]: I1127 10:23:23.002074 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:23Z","lastTransitionTime":"2025-11-27T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:23 crc kubenswrapper[4806]: I1127 10:23:23.105306 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:23 crc kubenswrapper[4806]: I1127 10:23:23.105355 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:23 crc kubenswrapper[4806]: I1127 10:23:23.105370 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:23 crc kubenswrapper[4806]: I1127 10:23:23.105389 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:23 crc kubenswrapper[4806]: I1127 10:23:23.105401 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:23Z","lastTransitionTime":"2025-11-27T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.115798 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:24 crc kubenswrapper[4806]: E1127 10:23:23.115991 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.208164 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.208217 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.208257 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.208280 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.208296 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:23Z","lastTransitionTime":"2025-11-27T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.311027 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.311079 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.311091 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.311110 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.311122 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:23Z","lastTransitionTime":"2025-11-27T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.413880 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.413922 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.413933 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.413951 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.413962 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:23Z","lastTransitionTime":"2025-11-27T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.517268 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.517299 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.517310 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.517325 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.517336 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:23Z","lastTransitionTime":"2025-11-27T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.620434 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.620471 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.620486 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.620508 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.620526 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:23Z","lastTransitionTime":"2025-11-27T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.723255 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.723283 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.723292 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.723309 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.723320 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:23Z","lastTransitionTime":"2025-11-27T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.825382 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.825416 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.825431 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.825453 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.825467 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:23Z","lastTransitionTime":"2025-11-27T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.927864 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.927888 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.927896 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.927908 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:23.927917 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:23Z","lastTransitionTime":"2025-11-27T10:23:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.031734 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.031763 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.031774 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.031789 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.031800 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.135551 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:24 crc kubenswrapper[4806]: E1127 10:23:24.135736 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.136007 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:24 crc kubenswrapper[4806]: E1127 10:23:24.136424 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.136679 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:24 crc kubenswrapper[4806]: E1127 10:23:24.136862 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.144976 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.148347 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.148440 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.148475 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.152383 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.257113 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.257164 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.257173 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.257192 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.257202 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.287177 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.287281 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.287295 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.287311 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.287322 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: E1127 10:23:24.304770 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.310595 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.310665 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.310679 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.310701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.310715 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: E1127 10:23:24.325936 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.330849 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.330895 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.330924 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.330942 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.330954 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: E1127 10:23:24.346932 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.353116 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.353184 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.353197 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.353220 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.353249 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: E1127 10:23:24.368049 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.373560 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.373617 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.373632 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.373653 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.373664 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: E1127 10:23:24.387727 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-27T10:23:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2d327c06-4eb5-4c2e-a11c-41971ef18290\\\",\\\"systemUUID\\\":\\\"0eb6d203-bd43-4b93-a96a-7190a2d0a1cc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-27T10:23:24Z is after 2025-08-24T17:21:41Z" Nov 27 10:23:24 crc kubenswrapper[4806]: E1127 10:23:24.387867 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.390415 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.390452 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.390463 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.390486 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.390499 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.494342 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.494394 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.494404 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.494424 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.494438 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.597197 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.597267 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.597283 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.597305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.597320 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.699987 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.700037 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.700050 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.700072 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.700119 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.803487 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.803549 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.803562 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.803582 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.803595 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.905893 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.905956 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.905968 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.905988 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:24 crc kubenswrapper[4806]: I1127 10:23:24.906003 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:24Z","lastTransitionTime":"2025-11-27T10:23:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.009575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.009637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.009648 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.009667 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.009678 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:25Z","lastTransitionTime":"2025-11-27T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.113179 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.113251 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.113265 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.113290 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.113305 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:25Z","lastTransitionTime":"2025-11-27T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.116676 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:25 crc kubenswrapper[4806]: E1127 10:23:25.117486 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.216945 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.216985 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.216996 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.217011 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.217023 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:25Z","lastTransitionTime":"2025-11-27T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.319866 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.319924 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.319941 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.319968 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.319986 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:25Z","lastTransitionTime":"2025-11-27T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.422344 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.422391 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.422404 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.422422 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.422436 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:25Z","lastTransitionTime":"2025-11-27T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.527647 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.527691 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.527701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.527718 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.527731 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:25Z","lastTransitionTime":"2025-11-27T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.629857 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.629900 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.629910 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.629922 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.629932 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:25Z","lastTransitionTime":"2025-11-27T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.733679 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.733729 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.733741 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.733758 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.733771 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:25Z","lastTransitionTime":"2025-11-27T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.836594 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.836649 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.836661 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.836679 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.836695 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:25Z","lastTransitionTime":"2025-11-27T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.940559 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.940638 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.940651 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.940671 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:25 crc kubenswrapper[4806]: I1127 10:23:25.940684 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:25Z","lastTransitionTime":"2025-11-27T10:23:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.044638 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.044693 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.044706 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.044728 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.044742 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:26Z","lastTransitionTime":"2025-11-27T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.118689 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:26 crc kubenswrapper[4806]: E1127 10:23:26.118958 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.118978 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:26 crc kubenswrapper[4806]: E1127 10:23:26.119206 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.118591 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:26 crc kubenswrapper[4806]: E1127 10:23:26.119905 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.150660 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.150803 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.150820 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.150841 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.150905 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:26Z","lastTransitionTime":"2025-11-27T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.164080 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=91.164053816 podStartE2EDuration="1m31.164053816s" podCreationTimestamp="2025-11-27 10:21:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:26.163503381 +0000 UTC m=+110.750094145" watchObservedRunningTime="2025-11-27 10:23:26.164053816 +0000 UTC m=+110.750644620" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.216019 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=88.215985764 podStartE2EDuration="1m28.215985764s" podCreationTimestamp="2025-11-27 10:21:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:26.215900461 +0000 UTC m=+110.802491235" watchObservedRunningTime="2025-11-27 10:23:26.215985764 +0000 UTC m=+110.802576518" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.232962 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-68rwq" podStartSLOduration=90.232935063 podStartE2EDuration="1m30.232935063s" podCreationTimestamp="2025-11-27 10:21:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:26.232889582 +0000 UTC m=+110.819480366" watchObservedRunningTime="2025-11-27 10:23:26.232935063 +0000 UTC m=+110.819525827" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.253767 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.253811 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.253825 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.253842 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.253854 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:26Z","lastTransitionTime":"2025-11-27T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.304593 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lqb5n" podStartSLOduration=89.304569058 podStartE2EDuration="1m29.304569058s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:26.285323915 +0000 UTC m=+110.871914689" watchObservedRunningTime="2025-11-27 10:23:26.304569058 +0000 UTC m=+110.891159822" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.304839 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=61.304831395 podStartE2EDuration="1m1.304831395s" podCreationTimestamp="2025-11-27 10:22:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:26.303443047 +0000 UTC m=+110.890033811" watchObservedRunningTime="2025-11-27 10:23:26.304831395 +0000 UTC m=+110.891422149" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.356846 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.356899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.356910 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.356933 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.356953 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:26Z","lastTransitionTime":"2025-11-27T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.359946 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podStartSLOduration=89.359921521 podStartE2EDuration="1m29.359921521s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:26.358926373 +0000 UTC m=+110.945517127" watchObservedRunningTime="2025-11-27 10:23:26.359921521 +0000 UTC m=+110.946512285" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.384974 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-44ndp" podStartSLOduration=89.384949955 podStartE2EDuration="1m29.384949955s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:26.384348768 +0000 UTC m=+110.970939532" watchObservedRunningTime="2025-11-27 10:23:26.384949955 +0000 UTC m=+110.971540719" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.401217 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-dzlc2" podStartSLOduration=90.401191195 podStartE2EDuration="1m30.401191195s" podCreationTimestamp="2025-11-27 10:21:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:26.40032484 +0000 UTC m=+110.986915604" watchObservedRunningTime="2025-11-27 10:23:26.401191195 +0000 UTC m=+110.987781959" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.453662 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=88.453636107 podStartE2EDuration="1m28.453636107s" podCreationTimestamp="2025-11-27 10:21:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:26.451273761 +0000 UTC m=+111.037864535" watchObservedRunningTime="2025-11-27 10:23:26.453636107 +0000 UTC m=+111.040226871" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.453926 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=41.453919654 podStartE2EDuration="41.453919654s" podCreationTimestamp="2025-11-27 10:22:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:26.42415079 +0000 UTC m=+111.010741544" watchObservedRunningTime="2025-11-27 10:23:26.453919654 +0000 UTC m=+111.040510418" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.460674 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.460717 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.460730 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.460798 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.460827 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:26Z","lastTransitionTime":"2025-11-27T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.529665 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-5cx6g" podStartSLOduration=89.529627051 podStartE2EDuration="1m29.529627051s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:26.528278505 +0000 UTC m=+111.114869269" watchObservedRunningTime="2025-11-27 10:23:26.529627051 +0000 UTC m=+111.116217815" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.564635 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.564722 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.564737 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.564764 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.564778 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:26Z","lastTransitionTime":"2025-11-27T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.667942 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.667994 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.668006 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.668026 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.668038 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:26Z","lastTransitionTime":"2025-11-27T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.772159 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.772830 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.773050 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.773446 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.773593 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:26Z","lastTransitionTime":"2025-11-27T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.876979 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.877030 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.877051 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.877109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.877124 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:26Z","lastTransitionTime":"2025-11-27T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.980318 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.980390 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.980412 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.980439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:26 crc kubenswrapper[4806]: I1127 10:23:26.980460 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:26Z","lastTransitionTime":"2025-11-27T10:23:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.083546 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.083644 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.083658 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.083683 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.083697 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:27Z","lastTransitionTime":"2025-11-27T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.116853 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:27 crc kubenswrapper[4806]: E1127 10:23:27.117214 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.187696 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.187904 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.187931 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.187953 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.187965 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:27Z","lastTransitionTime":"2025-11-27T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.291601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.291659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.291670 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.291692 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.291705 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:27Z","lastTransitionTime":"2025-11-27T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.394172 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.394222 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.394249 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.394269 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.394285 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:27Z","lastTransitionTime":"2025-11-27T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.497816 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.497924 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.497942 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.497968 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.497985 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:27Z","lastTransitionTime":"2025-11-27T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.600763 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.600845 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.600882 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.600903 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.600918 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:27Z","lastTransitionTime":"2025-11-27T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.704206 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.704300 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.704313 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.704337 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.704353 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:27Z","lastTransitionTime":"2025-11-27T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.808112 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.808177 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.808197 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.808220 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.808261 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:27Z","lastTransitionTime":"2025-11-27T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.912044 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.912104 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.912121 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.912146 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:27 crc kubenswrapper[4806]: I1127 10:23:27.912164 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:27Z","lastTransitionTime":"2025-11-27T10:23:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.016473 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.016568 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.016580 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.016602 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.016613 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:28Z","lastTransitionTime":"2025-11-27T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.116261 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.116330 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.116686 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:28 crc kubenswrapper[4806]: E1127 10:23:28.116955 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:28 crc kubenswrapper[4806]: E1127 10:23:28.117079 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:28 crc kubenswrapper[4806]: E1127 10:23:28.117291 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.118165 4806 scope.go:117] "RemoveContainer" containerID="faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084" Nov 27 10:23:28 crc kubenswrapper[4806]: E1127 10:23:28.118391 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rsgmm_openshift-ovn-kubernetes(e8ed6bbc-965a-412c-a3ee-4763035aa34e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.120198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.120261 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.120276 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.120298 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.120310 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:28Z","lastTransitionTime":"2025-11-27T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.223574 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.223625 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.223637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.223657 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.223670 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:28Z","lastTransitionTime":"2025-11-27T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.326061 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.326116 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.326127 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.326143 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.326153 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:28Z","lastTransitionTime":"2025-11-27T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.429024 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.429083 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.429094 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.429116 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.429130 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:28Z","lastTransitionTime":"2025-11-27T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.532562 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.532637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.532650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.532681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.532697 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:28Z","lastTransitionTime":"2025-11-27T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.635339 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.635401 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.635414 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.635446 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.635463 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:28Z","lastTransitionTime":"2025-11-27T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.743404 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.743460 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.743477 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.743497 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.743509 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:28Z","lastTransitionTime":"2025-11-27T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.847561 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.847631 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.847648 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.847672 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.847690 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:28Z","lastTransitionTime":"2025-11-27T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.951043 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.951092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.951104 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.951122 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:28 crc kubenswrapper[4806]: I1127 10:23:28.951135 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:28Z","lastTransitionTime":"2025-11-27T10:23:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.055154 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.055200 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.055213 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.055260 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.055273 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:29Z","lastTransitionTime":"2025-11-27T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.116555 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:29 crc kubenswrapper[4806]: E1127 10:23:29.116805 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.158903 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.158953 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.158962 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.158979 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.158989 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:29Z","lastTransitionTime":"2025-11-27T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.263301 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.263362 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.263376 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.263401 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.263415 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:29Z","lastTransitionTime":"2025-11-27T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.366613 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.366669 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.366683 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.366707 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.366723 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:29Z","lastTransitionTime":"2025-11-27T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.470398 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.470452 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.470463 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.470483 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.470495 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:29Z","lastTransitionTime":"2025-11-27T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.572904 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.572953 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.572965 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.572984 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.572994 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:29Z","lastTransitionTime":"2025-11-27T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.676193 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.676255 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.676265 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.676282 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.676296 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:29Z","lastTransitionTime":"2025-11-27T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.778549 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.778632 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.778645 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.778661 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.778681 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:29Z","lastTransitionTime":"2025-11-27T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.881944 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.882005 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.882017 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.882034 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.882045 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:29Z","lastTransitionTime":"2025-11-27T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.984956 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.984995 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.985004 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.985022 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:29 crc kubenswrapper[4806]: I1127 10:23:29.985032 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:29Z","lastTransitionTime":"2025-11-27T10:23:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.088864 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.088946 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.088987 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.089012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.089027 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:30Z","lastTransitionTime":"2025-11-27T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.116314 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.116350 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.116394 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:30 crc kubenswrapper[4806]: E1127 10:23:30.116493 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:30 crc kubenswrapper[4806]: E1127 10:23:30.116617 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:30 crc kubenswrapper[4806]: E1127 10:23:30.116814 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.192027 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.192092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.192107 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.192126 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.192537 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:30Z","lastTransitionTime":"2025-11-27T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.295551 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.295612 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.295629 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.295655 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.295673 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:30Z","lastTransitionTime":"2025-11-27T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.398652 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.398706 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.398720 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.398744 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.398759 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:30Z","lastTransitionTime":"2025-11-27T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.501705 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.501905 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.501922 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.502005 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.502021 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:30Z","lastTransitionTime":"2025-11-27T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.605066 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.605119 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.605134 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.605152 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.605165 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:30Z","lastTransitionTime":"2025-11-27T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.708289 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.708370 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.708385 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.708415 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.708430 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:30Z","lastTransitionTime":"2025-11-27T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.811343 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.811721 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.811785 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.811854 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.811923 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:30Z","lastTransitionTime":"2025-11-27T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.914206 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.914261 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.914272 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.914289 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:30 crc kubenswrapper[4806]: I1127 10:23:30.914301 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:30Z","lastTransitionTime":"2025-11-27T10:23:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.016748 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.016804 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.016817 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.016840 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.016869 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:31Z","lastTransitionTime":"2025-11-27T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.116248 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:31 crc kubenswrapper[4806]: E1127 10:23:31.116476 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.121046 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.121106 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.121119 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.121139 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.121156 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:31Z","lastTransitionTime":"2025-11-27T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.224377 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.224426 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.224443 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.224464 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.224481 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:31Z","lastTransitionTime":"2025-11-27T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.327852 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.327908 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.327930 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.327952 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.327963 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:31Z","lastTransitionTime":"2025-11-27T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.431293 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.431348 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.431364 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.431387 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.431408 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:31Z","lastTransitionTime":"2025-11-27T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.534562 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.534623 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.534636 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.534656 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.534669 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:31Z","lastTransitionTime":"2025-11-27T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.637516 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.637581 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.637601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.637627 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.637648 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:31Z","lastTransitionTime":"2025-11-27T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.741678 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.741976 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.742055 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.742094 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.742478 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:31Z","lastTransitionTime":"2025-11-27T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.760298 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5cx6g_39930cf4-5f3c-42cc-9384-d366bb1d6a78/kube-multus/1.log" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.761532 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5cx6g_39930cf4-5f3c-42cc-9384-d366bb1d6a78/kube-multus/0.log" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.761626 4806 generic.go:334] "Generic (PLEG): container finished" podID="39930cf4-5f3c-42cc-9384-d366bb1d6a78" containerID="04e5f8cbf4179151c6868e9168364cc8fa754b47829422ce6a64fa25edf7a949" exitCode=1 Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.761673 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5cx6g" event={"ID":"39930cf4-5f3c-42cc-9384-d366bb1d6a78","Type":"ContainerDied","Data":"04e5f8cbf4179151c6868e9168364cc8fa754b47829422ce6a64fa25edf7a949"} Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.761720 4806 scope.go:117] "RemoveContainer" containerID="6dfdcfa82e3763407b41ce9193811b5ffc9e8f5747085166209836e0ba8ccfc2" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.762449 4806 scope.go:117] "RemoveContainer" containerID="04e5f8cbf4179151c6868e9168364cc8fa754b47829422ce6a64fa25edf7a949" Nov 27 10:23:31 crc kubenswrapper[4806]: E1127 10:23:31.762730 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-5cx6g_openshift-multus(39930cf4-5f3c-42cc-9384-d366bb1d6a78)\"" pod="openshift-multus/multus-5cx6g" podUID="39930cf4-5f3c-42cc-9384-d366bb1d6a78" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.855133 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.855744 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.855760 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.855786 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.855803 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:31Z","lastTransitionTime":"2025-11-27T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.958751 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.958813 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.958826 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.958849 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:31 crc kubenswrapper[4806]: I1127 10:23:31.958868 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:31Z","lastTransitionTime":"2025-11-27T10:23:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.062307 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.062349 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.062357 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.062372 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.062382 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:32Z","lastTransitionTime":"2025-11-27T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.115982 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:32 crc kubenswrapper[4806]: E1127 10:23:32.116123 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.116446 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.116555 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:32 crc kubenswrapper[4806]: E1127 10:23:32.116661 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:32 crc kubenswrapper[4806]: E1127 10:23:32.116745 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.165314 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.165363 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.165375 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.165394 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.165406 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:32Z","lastTransitionTime":"2025-11-27T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.268956 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.269013 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.269025 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.269044 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.269055 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:32Z","lastTransitionTime":"2025-11-27T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.372090 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.372138 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.372150 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.372169 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.372182 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:32Z","lastTransitionTime":"2025-11-27T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.474799 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.474878 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.474899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.474929 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.474947 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:32Z","lastTransitionTime":"2025-11-27T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.578537 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.578587 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.578603 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.578622 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.578636 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:32Z","lastTransitionTime":"2025-11-27T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.681838 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.681900 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.681911 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.681934 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.681948 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:32Z","lastTransitionTime":"2025-11-27T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.778357 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5cx6g_39930cf4-5f3c-42cc-9384-d366bb1d6a78/kube-multus/1.log" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.785216 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.785379 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.785396 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.785460 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.785478 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:32Z","lastTransitionTime":"2025-11-27T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.888860 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.888903 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.888912 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.888927 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.888937 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:32Z","lastTransitionTime":"2025-11-27T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.991593 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.991647 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.991660 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.991673 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:32 crc kubenswrapper[4806]: I1127 10:23:32.991682 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:32Z","lastTransitionTime":"2025-11-27T10:23:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.094316 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.094388 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.094406 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.094431 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.094452 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:33Z","lastTransitionTime":"2025-11-27T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.116163 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:33 crc kubenswrapper[4806]: E1127 10:23:33.116489 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.196732 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.196780 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.196794 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.196813 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.196828 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:33Z","lastTransitionTime":"2025-11-27T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.299761 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.299869 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.299939 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.299977 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.300050 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:33Z","lastTransitionTime":"2025-11-27T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.403126 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.403223 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.403239 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.403254 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.403336 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:33Z","lastTransitionTime":"2025-11-27T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.506116 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.506160 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.506210 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.506281 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.506301 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:33Z","lastTransitionTime":"2025-11-27T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.608596 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.608644 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.608706 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.608728 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.608745 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:33Z","lastTransitionTime":"2025-11-27T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.712003 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.712076 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.712089 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.712107 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.712119 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:33Z","lastTransitionTime":"2025-11-27T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.814396 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.814527 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.814558 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.814591 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.814614 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:33Z","lastTransitionTime":"2025-11-27T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.917666 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.917738 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.917762 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.917794 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:33 crc kubenswrapper[4806]: I1127 10:23:33.917815 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:33Z","lastTransitionTime":"2025-11-27T10:23:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.020885 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.020954 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.020965 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.020986 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.021007 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:34Z","lastTransitionTime":"2025-11-27T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.116279 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.116344 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.116445 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:34 crc kubenswrapper[4806]: E1127 10:23:34.116491 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:34 crc kubenswrapper[4806]: E1127 10:23:34.116626 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:34 crc kubenswrapper[4806]: E1127 10:23:34.116778 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.128162 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.128228 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.128280 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.128302 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.128317 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:34Z","lastTransitionTime":"2025-11-27T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.232658 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.232722 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.232742 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.232773 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.232793 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:34Z","lastTransitionTime":"2025-11-27T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.335055 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.335092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.335102 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.335115 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.335126 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:34Z","lastTransitionTime":"2025-11-27T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.438326 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.438389 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.438408 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.438433 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.438453 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:34Z","lastTransitionTime":"2025-11-27T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.474616 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.474666 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.474681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.474700 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.474716 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-27T10:23:34Z","lastTransitionTime":"2025-11-27T10:23:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.549037 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx"] Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.549829 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.553522 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.553634 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.554011 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.554297 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.630708 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7497aa45-5003-4796-b315-ad8a276b060e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.630776 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7497aa45-5003-4796-b315-ad8a276b060e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.630843 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7497aa45-5003-4796-b315-ad8a276b060e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.631047 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7497aa45-5003-4796-b315-ad8a276b060e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.631335 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7497aa45-5003-4796-b315-ad8a276b060e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.733164 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7497aa45-5003-4796-b315-ad8a276b060e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.733556 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7497aa45-5003-4796-b315-ad8a276b060e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.733648 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7497aa45-5003-4796-b315-ad8a276b060e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.733690 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7497aa45-5003-4796-b315-ad8a276b060e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.733795 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7497aa45-5003-4796-b315-ad8a276b060e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.733843 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7497aa45-5003-4796-b315-ad8a276b060e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.733952 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7497aa45-5003-4796-b315-ad8a276b060e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.734336 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7497aa45-5003-4796-b315-ad8a276b060e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.745608 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7497aa45-5003-4796-b315-ad8a276b060e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.754924 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7497aa45-5003-4796-b315-ad8a276b060e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mpvkx\" (UID: \"7497aa45-5003-4796-b315-ad8a276b060e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:34 crc kubenswrapper[4806]: I1127 10:23:34.866542 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" Nov 27 10:23:35 crc kubenswrapper[4806]: I1127 10:23:35.116589 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:35 crc kubenswrapper[4806]: E1127 10:23:35.116737 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:35 crc kubenswrapper[4806]: I1127 10:23:35.788649 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" event={"ID":"7497aa45-5003-4796-b315-ad8a276b060e","Type":"ContainerStarted","Data":"b3de742732a5961f0c6ceb4ef4b6d21ba6869c9be61c4097a8265d6a3aec7435"} Nov 27 10:23:35 crc kubenswrapper[4806]: I1127 10:23:35.788694 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" event={"ID":"7497aa45-5003-4796-b315-ad8a276b060e","Type":"ContainerStarted","Data":"d7bf1c4fad83e477f1cd98261bcdb994a1c80e7b7b5e2289a12c95b4277b5576"} Nov 27 10:23:35 crc kubenswrapper[4806]: I1127 10:23:35.811432 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mpvkx" podStartSLOduration=98.811418586 podStartE2EDuration="1m38.811418586s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:35.80977219 +0000 UTC m=+120.396362954" watchObservedRunningTime="2025-11-27 10:23:35.811418586 +0000 UTC m=+120.398009350" Nov 27 10:23:36 crc kubenswrapper[4806]: E1127 10:23:36.088788 4806 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 27 10:23:36 crc kubenswrapper[4806]: I1127 10:23:36.117165 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:36 crc kubenswrapper[4806]: I1127 10:23:36.117246 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:36 crc kubenswrapper[4806]: I1127 10:23:36.119515 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:36 crc kubenswrapper[4806]: E1127 10:23:36.119527 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:36 crc kubenswrapper[4806]: E1127 10:23:36.119622 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:36 crc kubenswrapper[4806]: E1127 10:23:36.119692 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:36 crc kubenswrapper[4806]: E1127 10:23:36.231946 4806 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 10:23:37 crc kubenswrapper[4806]: I1127 10:23:37.116496 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:37 crc kubenswrapper[4806]: E1127 10:23:37.116620 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:38 crc kubenswrapper[4806]: I1127 10:23:38.115883 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:38 crc kubenswrapper[4806]: I1127 10:23:38.115884 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:38 crc kubenswrapper[4806]: E1127 10:23:38.116021 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:38 crc kubenswrapper[4806]: E1127 10:23:38.116175 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:38 crc kubenswrapper[4806]: I1127 10:23:38.116219 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:38 crc kubenswrapper[4806]: E1127 10:23:38.116320 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:39 crc kubenswrapper[4806]: I1127 10:23:39.115723 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:39 crc kubenswrapper[4806]: E1127 10:23:39.116674 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:39 crc kubenswrapper[4806]: I1127 10:23:39.125037 4806 scope.go:117] "RemoveContainer" containerID="faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084" Nov 27 10:23:39 crc kubenswrapper[4806]: I1127 10:23:39.803016 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/3.log" Nov 27 10:23:39 crc kubenswrapper[4806]: I1127 10:23:39.805099 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerStarted","Data":"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b"} Nov 27 10:23:39 crc kubenswrapper[4806]: I1127 10:23:39.806135 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:23:39 crc kubenswrapper[4806]: I1127 10:23:39.840176 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podStartSLOduration=102.840162709 podStartE2EDuration="1m42.840162709s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:39.838471002 +0000 UTC m=+124.425061766" watchObservedRunningTime="2025-11-27 10:23:39.840162709 +0000 UTC m=+124.426753473" Nov 27 10:23:40 crc kubenswrapper[4806]: I1127 10:23:40.023413 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-p2gpw"] Nov 27 10:23:40 crc kubenswrapper[4806]: I1127 10:23:40.023522 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:40 crc kubenswrapper[4806]: E1127 10:23:40.023613 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:40 crc kubenswrapper[4806]: I1127 10:23:40.116139 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:40 crc kubenswrapper[4806]: I1127 10:23:40.116140 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:40 crc kubenswrapper[4806]: E1127 10:23:40.116340 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:40 crc kubenswrapper[4806]: E1127 10:23:40.116354 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:40 crc kubenswrapper[4806]: I1127 10:23:40.116672 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:40 crc kubenswrapper[4806]: E1127 10:23:40.116826 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:41 crc kubenswrapper[4806]: I1127 10:23:41.115602 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:41 crc kubenswrapper[4806]: E1127 10:23:41.115824 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:41 crc kubenswrapper[4806]: E1127 10:23:41.233697 4806 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 27 10:23:42 crc kubenswrapper[4806]: I1127 10:23:42.115757 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:42 crc kubenswrapper[4806]: I1127 10:23:42.115865 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:42 crc kubenswrapper[4806]: I1127 10:23:42.115890 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:42 crc kubenswrapper[4806]: E1127 10:23:42.116223 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:42 crc kubenswrapper[4806]: E1127 10:23:42.116387 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:42 crc kubenswrapper[4806]: E1127 10:23:42.116605 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:43 crc kubenswrapper[4806]: I1127 10:23:43.116158 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:43 crc kubenswrapper[4806]: E1127 10:23:43.116391 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:44 crc kubenswrapper[4806]: I1127 10:23:44.116423 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:44 crc kubenswrapper[4806]: I1127 10:23:44.116568 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:44 crc kubenswrapper[4806]: E1127 10:23:44.116575 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:44 crc kubenswrapper[4806]: E1127 10:23:44.116635 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:44 crc kubenswrapper[4806]: I1127 10:23:44.116654 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:44 crc kubenswrapper[4806]: E1127 10:23:44.116890 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:44 crc kubenswrapper[4806]: I1127 10:23:44.117059 4806 scope.go:117] "RemoveContainer" containerID="04e5f8cbf4179151c6868e9168364cc8fa754b47829422ce6a64fa25edf7a949" Nov 27 10:23:44 crc kubenswrapper[4806]: I1127 10:23:44.827921 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5cx6g_39930cf4-5f3c-42cc-9384-d366bb1d6a78/kube-multus/1.log" Nov 27 10:23:44 crc kubenswrapper[4806]: I1127 10:23:44.827976 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5cx6g" event={"ID":"39930cf4-5f3c-42cc-9384-d366bb1d6a78","Type":"ContainerStarted","Data":"6e8529945bfb913b81082452411a2bfcf1b07dde94eaa1f2edce41149edb767e"} Nov 27 10:23:45 crc kubenswrapper[4806]: I1127 10:23:45.116204 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:45 crc kubenswrapper[4806]: E1127 10:23:45.116513 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p2gpw" podUID="3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e" Nov 27 10:23:46 crc kubenswrapper[4806]: I1127 10:23:46.115793 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:46 crc kubenswrapper[4806]: I1127 10:23:46.115908 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:46 crc kubenswrapper[4806]: I1127 10:23:46.116360 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:46 crc kubenswrapper[4806]: E1127 10:23:46.118919 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 27 10:23:46 crc kubenswrapper[4806]: E1127 10:23:46.119059 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 27 10:23:46 crc kubenswrapper[4806]: E1127 10:23:46.119177 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 27 10:23:47 crc kubenswrapper[4806]: I1127 10:23:47.116537 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:23:47 crc kubenswrapper[4806]: I1127 10:23:47.120042 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 27 10:23:47 crc kubenswrapper[4806]: I1127 10:23:47.120382 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 27 10:23:48 crc kubenswrapper[4806]: I1127 10:23:48.116572 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:23:48 crc kubenswrapper[4806]: I1127 10:23:48.116620 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:23:48 crc kubenswrapper[4806]: I1127 10:23:48.116653 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:23:48 crc kubenswrapper[4806]: I1127 10:23:48.120421 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 27 10:23:48 crc kubenswrapper[4806]: I1127 10:23:48.120428 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 27 10:23:48 crc kubenswrapper[4806]: I1127 10:23:48.120510 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 27 10:23:48 crc kubenswrapper[4806]: I1127 10:23:48.122276 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.702552 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.745665 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-z4ndr"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.746042 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.749742 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.750297 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.750755 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pppcb"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.752074 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.755325 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8jh8g"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.758926 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.759698 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.760073 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.758940 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.762860 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zd4hv"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.759115 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.768952 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.790746 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.791164 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.791244 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.791377 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.791585 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.791853 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.791586 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.794889 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.796351 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-9ppfs"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.796827 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9ppfs" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.797580 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.797862 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.798080 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.798201 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.798321 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.798427 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.798550 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.798666 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.807510 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.807734 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.807538 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.808115 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.808185 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.808353 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.807602 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.807652 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.808027 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.808707 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.808825 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.808919 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.808071 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.809056 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.809088 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.808667 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.809262 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.809388 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.809408 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.809270 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.809620 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.809706 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.809858 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.810128 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.810259 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.810476 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.810676 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.810898 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.811567 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.812791 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.819088 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.819515 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.819610 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.819539 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.819805 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.821108 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.823702 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.826312 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-42h4s"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.826834 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.827327 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.828267 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.836715 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l956q"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.838083 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.843446 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.843932 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-6fqzk"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.844358 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.844772 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.846423 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.846506 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.846725 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.846884 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.846978 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.847064 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.847119 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.847144 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.847330 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.847514 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.847784 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.848286 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2szrd"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.848794 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.849449 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.849659 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.849834 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.850012 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.850172 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.850334 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.850453 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.850569 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.850729 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.850898 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.851065 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.851223 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.851395 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.856537 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.858626 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.860318 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.862395 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kmbdt"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.863173 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.872850 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.876438 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.877714 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.877960 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.882328 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.882751 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.883160 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.878417 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.880391 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.881581 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.883556 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.883571 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.884148 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.884800 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.884886 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.902995 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.904348 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8jvb8"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.904740 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.905665 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.905665 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.906073 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.906202 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.907411 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.910392 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.910775 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.911385 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.911672 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.911850 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.912455 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.914938 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.915146 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.916256 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.923988 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.928555 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.918113 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lzwl4"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.929172 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.929493 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.929519 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.929802 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.929946 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.930009 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-lzwl4" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.930136 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.930928 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.934209 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.934700 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-r7wlb"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.935157 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-r7wlb" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.934705 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.935175 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944462 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52480926-8957-4c79-a384-3e2d27f349af-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8gdsx\" (UID: \"52480926-8957-4c79-a384-3e2d27f349af\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944516 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944541 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93b844fb-455e-4a0a-9d75-52f1278df25c-service-ca-bundle\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944573 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-image-import-ca\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944603 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dqjqc\" (UID: \"6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944623 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944644 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a52b3a70-665d-49e5-a842-14ade4dcff1f-audit-dir\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944667 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjm5h\" (UniqueName: \"kubernetes.io/projected/a52b3a70-665d-49e5-a842-14ade4dcff1f-kube-api-access-jjm5h\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944697 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/47d1375e-f8bc-4406-8692-fe9aecb0e394-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944734 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/325e82c6-eacf-4180-bc5f-9334536ebe16-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zd4hv\" (UID: \"325e82c6-eacf-4180-bc5f-9334536ebe16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944767 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944792 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/325e82c6-eacf-4180-bc5f-9334536ebe16-images\") pod \"machine-api-operator-5694c8668f-zd4hv\" (UID: \"325e82c6-eacf-4180-bc5f-9334536ebe16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944836 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcqvp\" (UniqueName: \"kubernetes.io/projected/52480926-8957-4c79-a384-3e2d27f349af-kube-api-access-wcqvp\") pod \"openshift-apiserver-operator-796bbdcf4f-8gdsx\" (UID: \"52480926-8957-4c79-a384-3e2d27f349af\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944875 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93b844fb-455e-4a0a-9d75-52f1278df25c-serving-cert\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944894 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944914 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a52b3a70-665d-49e5-a842-14ade4dcff1f-encryption-config\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944937 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d26adb0c-6bf0-49f1-806b-06d96e333fc4-config\") pod \"machine-approver-56656f9798-sqqv6\" (UID: \"d26adb0c-6bf0-49f1-806b-06d96e333fc4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944963 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65b0faa1-6054-43a4-ab15-e59791790337-serving-cert\") pod \"openshift-config-operator-7777fb866f-pppcb\" (UID: \"65b0faa1-6054-43a4-ab15-e59791790337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.944981 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d52f4ad9-009f-41cf-a779-2a6515a2fbb4-serving-cert\") pod \"console-operator-58897d9998-42h4s\" (UID: \"d52f4ad9-009f-41cf-a779-2a6515a2fbb4\") " pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945028 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a52b3a70-665d-49e5-a842-14ade4dcff1f-etcd-client\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945053 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945073 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-audit-policies\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945093 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/65b0faa1-6054-43a4-ab15-e59791790337-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pppcb\" (UID: \"65b0faa1-6054-43a4-ab15-e59791790337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945132 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52480926-8957-4c79-a384-3e2d27f349af-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8gdsx\" (UID: \"52480926-8957-4c79-a384-3e2d27f349af\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945152 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a52b3a70-665d-49e5-a842-14ade4dcff1f-serving-cert\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945178 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945200 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/47d1375e-f8bc-4406-8692-fe9aecb0e394-audit-policies\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945223 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dqjqc\" (UID: \"6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945264 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93b844fb-455e-4a0a-9d75-52f1278df25c-config\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945297 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qxs4\" (UniqueName: \"kubernetes.io/projected/93b844fb-455e-4a0a-9d75-52f1278df25c-kube-api-access-5qxs4\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945320 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a52b3a70-665d-49e5-a842-14ade4dcff1f-node-pullsecrets\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945338 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/47d1375e-f8bc-4406-8692-fe9aecb0e394-etcd-client\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945359 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d26adb0c-6bf0-49f1-806b-06d96e333fc4-auth-proxy-config\") pod \"machine-approver-56656f9798-sqqv6\" (UID: \"d26adb0c-6bf0-49f1-806b-06d96e333fc4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945379 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-etcd-serving-ca\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945397 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hbnm\" (UniqueName: \"kubernetes.io/projected/d52f4ad9-009f-41cf-a779-2a6515a2fbb4-kube-api-access-7hbnm\") pod \"console-operator-58897d9998-42h4s\" (UID: \"d52f4ad9-009f-41cf-a779-2a6515a2fbb4\") " pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945416 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d26adb0c-6bf0-49f1-806b-06d96e333fc4-machine-approver-tls\") pod \"machine-approver-56656f9798-sqqv6\" (UID: \"d26adb0c-6bf0-49f1-806b-06d96e333fc4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945444 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d1375e-f8bc-4406-8692-fe9aecb0e394-serving-cert\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945465 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47d1375e-f8bc-4406-8692-fe9aecb0e394-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945486 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/47d1375e-f8bc-4406-8692-fe9aecb0e394-encryption-config\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945505 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0639fe21-6b99-4fad-be7b-6bd998c3472c-audit-dir\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945524 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tphxc\" (UniqueName: \"kubernetes.io/projected/52dd6d6d-3a15-4467-9d82-37fd2fa5794b-kube-api-access-tphxc\") pod \"downloads-7954f5f757-9ppfs\" (UID: \"52dd6d6d-3a15-4467-9d82-37fd2fa5794b\") " pod="openshift-console/downloads-7954f5f757-9ppfs" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945547 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d52f4ad9-009f-41cf-a779-2a6515a2fbb4-trusted-ca\") pod \"console-operator-58897d9998-42h4s\" (UID: \"d52f4ad9-009f-41cf-a779-2a6515a2fbb4\") " pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945581 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945598 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82hcz\" (UniqueName: \"kubernetes.io/projected/325e82c6-eacf-4180-bc5f-9334536ebe16-kube-api-access-82hcz\") pod \"machine-api-operator-5694c8668f-zd4hv\" (UID: \"325e82c6-eacf-4180-bc5f-9334536ebe16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945618 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945652 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d52f4ad9-009f-41cf-a779-2a6515a2fbb4-config\") pod \"console-operator-58897d9998-42h4s\" (UID: \"d52f4ad9-009f-41cf-a779-2a6515a2fbb4\") " pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945669 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/47d1375e-f8bc-4406-8692-fe9aecb0e394-audit-dir\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945687 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3ce016c-a670-46a3-9254-03fea2c6568b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qgxx9\" (UID: \"a3ce016c-a670-46a3-9254-03fea2c6568b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945707 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqf8w\" (UniqueName: \"kubernetes.io/projected/a3ce016c-a670-46a3-9254-03fea2c6568b-kube-api-access-sqf8w\") pod \"cluster-samples-operator-665b6dd947-qgxx9\" (UID: \"a3ce016c-a670-46a3-9254-03fea2c6568b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945732 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx8kd\" (UniqueName: \"kubernetes.io/projected/65b0faa1-6054-43a4-ab15-e59791790337-kube-api-access-hx8kd\") pod \"openshift-config-operator-7777fb866f-pppcb\" (UID: \"65b0faa1-6054-43a4-ab15-e59791790337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945751 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-audit\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945775 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945794 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945812 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7qs6\" (UniqueName: \"kubernetes.io/projected/0639fe21-6b99-4fad-be7b-6bd998c3472c-kube-api-access-n7qs6\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945830 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945848 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93b844fb-455e-4a0a-9d75-52f1278df25c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945873 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-config\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945893 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qntsv\" (UniqueName: \"kubernetes.io/projected/6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe-kube-api-access-qntsv\") pod \"openshift-controller-manager-operator-756b6f6bc6-dqjqc\" (UID: \"6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945912 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945930 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x49vx\" (UniqueName: \"kubernetes.io/projected/d26adb0c-6bf0-49f1-806b-06d96e333fc4-kube-api-access-x49vx\") pod \"machine-approver-56656f9798-sqqv6\" (UID: \"d26adb0c-6bf0-49f1-806b-06d96e333fc4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945952 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/325e82c6-eacf-4180-bc5f-9334536ebe16-config\") pod \"machine-api-operator-5694c8668f-zd4hv\" (UID: \"325e82c6-eacf-4180-bc5f-9334536ebe16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.945972 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzkwd\" (UniqueName: \"kubernetes.io/projected/47d1375e-f8bc-4406-8692-fe9aecb0e394-kube-api-access-mzkwd\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.948167 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.950166 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-z4ndr"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.950208 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-kpjrm"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.951278 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.952504 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zd4hv"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.964763 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.965389 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.965742 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.965962 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.966411 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7t7px"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.967063 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.968084 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-62tks"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.968713 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-62tks" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.969179 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.971300 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.971673 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.972830 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.973169 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.976143 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nflhv"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.988256 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.991900 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.992422 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.992483 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-28rbb"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.992857 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.993156 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.993323 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.993590 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.993646 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-57tr4"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.993852 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.994004 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-zbhcm"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.994032 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.994111 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-57tr4" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.994670 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9ppfs"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.994696 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-42h4s"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.994736 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zbhcm" Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.996865 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pppcb"] Nov 27 10:23:54 crc kubenswrapper[4806]: I1127 10:23:54.996886 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.000122 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.001598 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.002671 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2szrd"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.009478 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.009729 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l956q"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.011297 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6fqzk"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.012630 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.013500 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.014618 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-62tks"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.014937 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.016441 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-r7wlb"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.017394 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.020429 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-gwfw8"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.021096 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.021126 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lzwl4"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.021275 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gwfw8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.024759 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.025799 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.026865 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.028877 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.031644 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.042132 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.043936 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8jvb8"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.045323 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.046818 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.047352 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13187e9e-9b78-41ce-a2ff-7837dfd8eca6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fx54r\" (UID: \"13187e9e-9b78-41ce-a2ff-7837dfd8eca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.047457 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-oauth-serving-cert\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.047539 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c225ab90-51ee-400f-8902-774ad6270322-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-678qm\" (UID: \"c225ab90-51ee-400f-8902-774ad6270322\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.047653 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0639fe21-6b99-4fad-be7b-6bd998c3472c-audit-dir\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.047743 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tphxc\" (UniqueName: \"kubernetes.io/projected/52dd6d6d-3a15-4467-9d82-37fd2fa5794b-kube-api-access-tphxc\") pod \"downloads-7954f5f757-9ppfs\" (UID: \"52dd6d6d-3a15-4467-9d82-37fd2fa5794b\") " pod="openshift-console/downloads-7954f5f757-9ppfs" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.047820 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f25c1d0d-db4f-415b-a7ea-4d1b98e84819-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwk9c\" (UID: \"f25c1d0d-db4f-415b-a7ea-4d1b98e84819\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.047878 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d52f4ad9-009f-41cf-a779-2a6515a2fbb4-trusted-ca\") pod \"console-operator-58897d9998-42h4s\" (UID: \"d52f4ad9-009f-41cf-a779-2a6515a2fbb4\") " pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.047912 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.047938 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82hcz\" (UniqueName: \"kubernetes.io/projected/325e82c6-eacf-4180-bc5f-9334536ebe16-kube-api-access-82hcz\") pod \"machine-api-operator-5694c8668f-zd4hv\" (UID: \"325e82c6-eacf-4180-bc5f-9334536ebe16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.047963 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.047988 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx9d2\" (UniqueName: \"kubernetes.io/projected/13187e9e-9b78-41ce-a2ff-7837dfd8eca6-kube-api-access-xx9d2\") pod \"kube-storage-version-migrator-operator-b67b599dd-fx54r\" (UID: \"13187e9e-9b78-41ce-a2ff-7837dfd8eca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048005 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06f03ec3-796d-442f-8adc-93aa22466426-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fszts\" (UID: \"06f03ec3-796d-442f-8adc-93aa22466426\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048037 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d52f4ad9-009f-41cf-a779-2a6515a2fbb4-config\") pod \"console-operator-58897d9998-42h4s\" (UID: \"d52f4ad9-009f-41cf-a779-2a6515a2fbb4\") " pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048055 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06f03ec3-796d-442f-8adc-93aa22466426-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fszts\" (UID: \"06f03ec3-796d-442f-8adc-93aa22466426\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048072 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngjxb\" (UniqueName: \"kubernetes.io/projected/e858c9d3-3f1c-47db-a0c8-f0efac1ed114-kube-api-access-ngjxb\") pod \"multus-admission-controller-857f4d67dd-r7wlb\" (UID: \"e858c9d3-3f1c-47db-a0c8-f0efac1ed114\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r7wlb" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048091 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-config\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048117 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx8kd\" (UniqueName: \"kubernetes.io/projected/65b0faa1-6054-43a4-ab15-e59791790337-kube-api-access-hx8kd\") pod \"openshift-config-operator-7777fb866f-pppcb\" (UID: \"65b0faa1-6054-43a4-ab15-e59791790337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048135 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/47d1375e-f8bc-4406-8692-fe9aecb0e394-audit-dir\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048152 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3ce016c-a670-46a3-9254-03fea2c6568b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qgxx9\" (UID: \"a3ce016c-a670-46a3-9254-03fea2c6568b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048170 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqf8w\" (UniqueName: \"kubernetes.io/projected/a3ce016c-a670-46a3-9254-03fea2c6568b-kube-api-access-sqf8w\") pod \"cluster-samples-operator-665b6dd947-qgxx9\" (UID: \"a3ce016c-a670-46a3-9254-03fea2c6568b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048188 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e60de411-4929-4a3a-a5ed-1071da7ed16a-serving-cert\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048208 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqnvq\" (UniqueName: \"kubernetes.io/projected/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-kube-api-access-hqnvq\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048265 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b5f5\" (UniqueName: \"kubernetes.io/projected/4d69e866-000e-4748-9953-4d5f007f2669-kube-api-access-4b5f5\") pod \"machine-config-controller-84d6567774-rnhfh\" (UID: \"4d69e866-000e-4748-9953-4d5f007f2669\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048278 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048287 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-audit\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048315 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/97042267-61ca-4864-a50a-60587abc7619-console-serving-cert\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048336 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-client-ca\") pod \"route-controller-manager-6576b87f9c-xb7gk\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048359 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f25c1d0d-db4f-415b-a7ea-4d1b98e84819-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwk9c\" (UID: \"f25c1d0d-db4f-415b-a7ea-4d1b98e84819\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048379 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048398 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048419 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7qs6\" (UniqueName: \"kubernetes.io/projected/0639fe21-6b99-4fad-be7b-6bd998c3472c-kube-api-access-n7qs6\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048441 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93b844fb-455e-4a0a-9d75-52f1278df25c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048457 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-config\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048477 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048497 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-config\") pod \"route-controller-manager-6576b87f9c-xb7gk\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048522 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qntsv\" (UniqueName: \"kubernetes.io/projected/6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe-kube-api-access-qntsv\") pod \"openshift-controller-manager-operator-756b6f6bc6-dqjqc\" (UID: \"6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048530 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0639fe21-6b99-4fad-be7b-6bd998c3472c-audit-dir\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048542 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048584 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x49vx\" (UniqueName: \"kubernetes.io/projected/d26adb0c-6bf0-49f1-806b-06d96e333fc4-kube-api-access-x49vx\") pod \"machine-approver-56656f9798-sqqv6\" (UID: \"d26adb0c-6bf0-49f1-806b-06d96e333fc4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048618 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e60de411-4929-4a3a-a5ed-1071da7ed16a-etcd-client\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048642 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-console-config\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048675 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/325e82c6-eacf-4180-bc5f-9334536ebe16-config\") pod \"machine-api-operator-5694c8668f-zd4hv\" (UID: \"325e82c6-eacf-4180-bc5f-9334536ebe16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048706 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzkwd\" (UniqueName: \"kubernetes.io/projected/47d1375e-f8bc-4406-8692-fe9aecb0e394-kube-api-access-mzkwd\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048736 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e60de411-4929-4a3a-a5ed-1071da7ed16a-etcd-service-ca\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048768 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52480926-8957-4c79-a384-3e2d27f349af-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8gdsx\" (UID: \"52480926-8957-4c79-a384-3e2d27f349af\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048794 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048822 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-trusted-ca-bundle\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048853 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93b844fb-455e-4a0a-9d75-52f1278df25c-service-ca-bundle\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048880 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-image-import-ca\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048906 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqslv\" (UniqueName: \"kubernetes.io/projected/f893e28e-fb92-45b2-8c43-05ff99bf7d32-kube-api-access-kqslv\") pod \"ingress-operator-5b745b69d9-gnrs8\" (UID: \"f893e28e-fb92-45b2-8c43-05ff99bf7d32\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048939 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dqjqc\" (UID: \"6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048969 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.048998 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c225ab90-51ee-400f-8902-774ad6270322-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-678qm\" (UID: \"c225ab90-51ee-400f-8902-774ad6270322\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049027 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a52b3a70-665d-49e5-a842-14ade4dcff1f-audit-dir\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049051 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjm5h\" (UniqueName: \"kubernetes.io/projected/a52b3a70-665d-49e5-a842-14ade4dcff1f-kube-api-access-jjm5h\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049077 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/47d1375e-f8bc-4406-8692-fe9aecb0e394-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049105 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/325e82c6-eacf-4180-bc5f-9334536ebe16-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zd4hv\" (UID: \"325e82c6-eacf-4180-bc5f-9334536ebe16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049131 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13187e9e-9b78-41ce-a2ff-7837dfd8eca6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fx54r\" (UID: \"13187e9e-9b78-41ce-a2ff-7837dfd8eca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049160 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049188 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/325e82c6-eacf-4180-bc5f-9334536ebe16-images\") pod \"machine-api-operator-5694c8668f-zd4hv\" (UID: \"325e82c6-eacf-4180-bc5f-9334536ebe16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049215 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e623ba7e-e936-4ab0-a825-6139b2b33fe4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kb5rn\" (UID: \"e623ba7e-e936-4ab0-a825-6139b2b33fe4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049263 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-client-ca\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049287 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-serving-cert\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049338 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93b844fb-455e-4a0a-9d75-52f1278df25c-serving-cert\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049368 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcqvp\" (UniqueName: \"kubernetes.io/projected/52480926-8957-4c79-a384-3e2d27f349af-kube-api-access-wcqvp\") pod \"openshift-apiserver-operator-796bbdcf4f-8gdsx\" (UID: \"52480926-8957-4c79-a384-3e2d27f349af\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049395 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krmjc\" (UniqueName: \"kubernetes.io/projected/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-kube-api-access-krmjc\") pod \"route-controller-manager-6576b87f9c-xb7gk\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049429 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049457 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a52b3a70-665d-49e5-a842-14ade4dcff1f-encryption-config\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049486 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65b0faa1-6054-43a4-ab15-e59791790337-serving-cert\") pod \"openshift-config-operator-7777fb866f-pppcb\" (UID: \"65b0faa1-6054-43a4-ab15-e59791790337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049598 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/325e82c6-eacf-4180-bc5f-9334536ebe16-config\") pod \"machine-api-operator-5694c8668f-zd4hv\" (UID: \"325e82c6-eacf-4180-bc5f-9334536ebe16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049509 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d52f4ad9-009f-41cf-a779-2a6515a2fbb4-serving-cert\") pod \"console-operator-58897d9998-42h4s\" (UID: \"d52f4ad9-009f-41cf-a779-2a6515a2fbb4\") " pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049914 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a52b3a70-665d-49e5-a842-14ade4dcff1f-etcd-client\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049949 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d26adb0c-6bf0-49f1-806b-06d96e333fc4-config\") pod \"machine-approver-56656f9798-sqqv6\" (UID: \"d26adb0c-6bf0-49f1-806b-06d96e333fc4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049978 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-service-ca\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.050068 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4d69e866-000e-4748-9953-4d5f007f2669-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rnhfh\" (UID: \"4d69e866-000e-4748-9953-4d5f007f2669\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.050108 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.050135 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f893e28e-fb92-45b2-8c43-05ff99bf7d32-trusted-ca\") pod \"ingress-operator-5b745b69d9-gnrs8\" (UID: \"f893e28e-fb92-45b2-8c43-05ff99bf7d32\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.050273 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e623ba7e-e936-4ab0-a825-6139b2b33fe4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kb5rn\" (UID: \"e623ba7e-e936-4ab0-a825-6139b2b33fe4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.050359 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e60de411-4929-4a3a-a5ed-1071da7ed16a-config\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.050435 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7dtd\" (UniqueName: \"kubernetes.io/projected/97042267-61ca-4864-a50a-60587abc7619-kube-api-access-j7dtd\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.050506 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f25c1d0d-db4f-415b-a7ea-4d1b98e84819-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwk9c\" (UID: \"f25c1d0d-db4f-415b-a7ea-4d1b98e84819\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.051328 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.051427 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06f03ec3-796d-442f-8adc-93aa22466426-config\") pod \"kube-controller-manager-operator-78b949d7b-fszts\" (UID: \"06f03ec3-796d-442f-8adc-93aa22466426\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.051504 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-audit-policies\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.051579 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e858c9d3-3f1c-47db-a0c8-f0efac1ed114-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-r7wlb\" (UID: \"e858c9d3-3f1c-47db-a0c8-f0efac1ed114\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r7wlb" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.051645 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9crtj\" (UniqueName: \"kubernetes.io/projected/c225ab90-51ee-400f-8902-774ad6270322-kube-api-access-9crtj\") pod \"cluster-image-registry-operator-dc59b4c8b-678qm\" (UID: \"c225ab90-51ee-400f-8902-774ad6270322\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.051718 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/65b0faa1-6054-43a4-ab15-e59791790337-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pppcb\" (UID: \"65b0faa1-6054-43a4-ab15-e59791790337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.051793 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52480926-8957-4c79-a384-3e2d27f349af-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8gdsx\" (UID: \"52480926-8957-4c79-a384-3e2d27f349af\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.052790 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/65b0faa1-6054-43a4-ab15-e59791790337-available-featuregates\") pod \"openshift-config-operator-7777fb866f-pppcb\" (UID: \"65b0faa1-6054-43a4-ab15-e59791790337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.054010 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a52b3a70-665d-49e5-a842-14ade4dcff1f-serving-cert\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.054054 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f893e28e-fb92-45b2-8c43-05ff99bf7d32-metrics-tls\") pod \"ingress-operator-5b745b69d9-gnrs8\" (UID: \"f893e28e-fb92-45b2-8c43-05ff99bf7d32\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.054080 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e60de411-4929-4a3a-a5ed-1071da7ed16a-etcd-ca\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.054103 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4d69e866-000e-4748-9953-4d5f007f2669-proxy-tls\") pod \"machine-config-controller-84d6567774-rnhfh\" (UID: \"4d69e866-000e-4748-9953-4d5f007f2669\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.054869 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-image-import-ca\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.055435 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52480926-8957-4c79-a384-3e2d27f349af-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8gdsx\" (UID: \"52480926-8957-4c79-a384-3e2d27f349af\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.055465 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a52b3a70-665d-49e5-a842-14ade4dcff1f-audit-dir\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.056076 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93b844fb-455e-4a0a-9d75-52f1278df25c-service-ca-bundle\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.056174 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/47d1375e-f8bc-4406-8692-fe9aecb0e394-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.056458 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.056507 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dqjqc\" (UID: \"6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.049397 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52480926-8957-4c79-a384-3e2d27f349af-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8gdsx\" (UID: \"52480926-8957-4c79-a384-3e2d27f349af\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.057923 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-dqjqc\" (UID: \"6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.058733 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/325e82c6-eacf-4180-bc5f-9334536ebe16-images\") pod \"machine-api-operator-5694c8668f-zd4hv\" (UID: \"325e82c6-eacf-4180-bc5f-9334536ebe16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.056537 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.058895 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.059078 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-audit-policies\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.059926 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.060273 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.060553 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a52b3a70-665d-49e5-a842-14ade4dcff1f-etcd-client\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.060808 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93b844fb-455e-4a0a-9d75-52f1278df25c-serving-cert\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.060980 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a52b3a70-665d-49e5-a842-14ade4dcff1f-encryption-config\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.061003 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-dqjqc\" (UID: \"6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.061455 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.062295 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.063049 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a52b3a70-665d-49e5-a842-14ade4dcff1f-serving-cert\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.063062 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d52f4ad9-009f-41cf-a779-2a6515a2fbb4-config\") pod \"console-operator-58897d9998-42h4s\" (UID: \"d52f4ad9-009f-41cf-a779-2a6515a2fbb4\") " pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.063922 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kmbdt"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.063960 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7t7px"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.063975 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8jh8g"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.064064 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65b0faa1-6054-43a4-ab15-e59791790337-serving-cert\") pod \"openshift-config-operator-7777fb866f-pppcb\" (UID: \"65b0faa1-6054-43a4-ab15-e59791790337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.064498 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/325e82c6-eacf-4180-bc5f-9334536ebe16-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zd4hv\" (UID: \"325e82c6-eacf-4180-bc5f-9334536ebe16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.064645 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/47d1375e-f8bc-4406-8692-fe9aecb0e394-audit-dir\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.065987 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.066353 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.066507 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.066544 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-audit\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.066721 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/47d1375e-f8bc-4406-8692-fe9aecb0e394-audit-policies\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.066749 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e623ba7e-e936-4ab0-a825-6139b2b33fe4-config\") pod \"kube-apiserver-operator-766d6c64bb-kb5rn\" (UID: \"e623ba7e-e936-4ab0-a825-6139b2b33fe4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.066767 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-serving-cert\") pod \"route-controller-manager-6576b87f9c-xb7gk\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.066798 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93b844fb-455e-4a0a-9d75-52f1278df25c-config\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.066880 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.067300 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-config\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.067364 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93b844fb-455e-4a0a-9d75-52f1278df25c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.067939 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-28rbb"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.068153 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d52f4ad9-009f-41cf-a779-2a6515a2fbb4-trusted-ca\") pod \"console-operator-58897d9998-42h4s\" (UID: \"d52f4ad9-009f-41cf-a779-2a6515a2fbb4\") " pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.068591 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/47d1375e-f8bc-4406-8692-fe9aecb0e394-audit-policies\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.068977 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93b844fb-455e-4a0a-9d75-52f1278df25c-config\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.069032 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qxs4\" (UniqueName: \"kubernetes.io/projected/93b844fb-455e-4a0a-9d75-52f1278df25c-kube-api-access-5qxs4\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.069069 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a52b3a70-665d-49e5-a842-14ade4dcff1f-node-pullsecrets\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.069092 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/47d1375e-f8bc-4406-8692-fe9aecb0e394-etcd-client\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.069133 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d26adb0c-6bf0-49f1-806b-06d96e333fc4-auth-proxy-config\") pod \"machine-approver-56656f9798-sqqv6\" (UID: \"d26adb0c-6bf0-49f1-806b-06d96e333fc4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.069155 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f893e28e-fb92-45b2-8c43-05ff99bf7d32-bound-sa-token\") pod \"ingress-operator-5b745b69d9-gnrs8\" (UID: \"f893e28e-fb92-45b2-8c43-05ff99bf7d32\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.069173 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/97042267-61ca-4864-a50a-60587abc7619-console-oauth-config\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.069751 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3ce016c-a670-46a3-9254-03fea2c6568b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qgxx9\" (UID: \"a3ce016c-a670-46a3-9254-03fea2c6568b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.069816 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zbhcm"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.070266 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hbnm\" (UniqueName: \"kubernetes.io/projected/d52f4ad9-009f-41cf-a779-2a6515a2fbb4-kube-api-access-7hbnm\") pod \"console-operator-58897d9998-42h4s\" (UID: \"d52f4ad9-009f-41cf-a779-2a6515a2fbb4\") " pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.070298 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-etcd-serving-ca\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.070356 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d1375e-f8bc-4406-8692-fe9aecb0e394-serving-cert\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.070374 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47d1375e-f8bc-4406-8692-fe9aecb0e394-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.070411 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/47d1375e-f8bc-4406-8692-fe9aecb0e394-encryption-config\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.070428 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d26adb0c-6bf0-49f1-806b-06d96e333fc4-machine-approver-tls\") pod \"machine-approver-56656f9798-sqqv6\" (UID: \"d26adb0c-6bf0-49f1-806b-06d96e333fc4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.070447 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fktpj\" (UniqueName: \"kubernetes.io/projected/e60de411-4929-4a3a-a5ed-1071da7ed16a-kube-api-access-fktpj\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.070467 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c225ab90-51ee-400f-8902-774ad6270322-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-678qm\" (UID: \"c225ab90-51ee-400f-8902-774ad6270322\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.070772 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nflhv"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.070809 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a52b3a70-665d-49e5-a842-14ade4dcff1f-node-pullsecrets\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.071643 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a52b3a70-665d-49e5-a842-14ade4dcff1f-etcd-serving-ca\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.071848 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47d1375e-f8bc-4406-8692-fe9aecb0e394-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.072213 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-57tr4"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.075410 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.075471 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.075833 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d26adb0c-6bf0-49f1-806b-06d96e333fc4-config\") pod \"machine-approver-56656f9798-sqqv6\" (UID: \"d26adb0c-6bf0-49f1-806b-06d96e333fc4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.077814 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d26adb0c-6bf0-49f1-806b-06d96e333fc4-auth-proxy-config\") pod \"machine-approver-56656f9798-sqqv6\" (UID: \"d26adb0c-6bf0-49f1-806b-06d96e333fc4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.077870 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.078888 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.080163 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z979k"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.081160 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z979k"] Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.081270 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.081798 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d26adb0c-6bf0-49f1-806b-06d96e333fc4-machine-approver-tls\") pod \"machine-approver-56656f9798-sqqv6\" (UID: \"d26adb0c-6bf0-49f1-806b-06d96e333fc4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.083427 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.083643 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d52f4ad9-009f-41cf-a779-2a6515a2fbb4-serving-cert\") pod \"console-operator-58897d9998-42h4s\" (UID: \"d52f4ad9-009f-41cf-a779-2a6515a2fbb4\") " pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.083808 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d1375e-f8bc-4406-8692-fe9aecb0e394-serving-cert\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.086246 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.086311 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/47d1375e-f8bc-4406-8692-fe9aecb0e394-etcd-client\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.086560 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/47d1375e-f8bc-4406-8692-fe9aecb0e394-encryption-config\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.088373 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.107948 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.128766 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.147824 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.167845 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171471 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/97042267-61ca-4864-a50a-60587abc7619-console-serving-cert\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171500 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-client-ca\") pod \"route-controller-manager-6576b87f9c-xb7gk\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171518 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f25c1d0d-db4f-415b-a7ea-4d1b98e84819-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwk9c\" (UID: \"f25c1d0d-db4f-415b-a7ea-4d1b98e84819\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171540 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-config\") pod \"route-controller-manager-6576b87f9c-xb7gk\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171572 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e60de411-4929-4a3a-a5ed-1071da7ed16a-etcd-client\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171613 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-console-config\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171634 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e60de411-4929-4a3a-a5ed-1071da7ed16a-etcd-service-ca\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171648 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-trusted-ca-bundle\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171666 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqslv\" (UniqueName: \"kubernetes.io/projected/f893e28e-fb92-45b2-8c43-05ff99bf7d32-kube-api-access-kqslv\") pod \"ingress-operator-5b745b69d9-gnrs8\" (UID: \"f893e28e-fb92-45b2-8c43-05ff99bf7d32\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171685 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c225ab90-51ee-400f-8902-774ad6270322-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-678qm\" (UID: \"c225ab90-51ee-400f-8902-774ad6270322\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171706 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13187e9e-9b78-41ce-a2ff-7837dfd8eca6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fx54r\" (UID: \"13187e9e-9b78-41ce-a2ff-7837dfd8eca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171720 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-client-ca\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171736 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-serving-cert\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171753 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e623ba7e-e936-4ab0-a825-6139b2b33fe4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kb5rn\" (UID: \"e623ba7e-e936-4ab0-a825-6139b2b33fe4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171773 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krmjc\" (UniqueName: \"kubernetes.io/projected/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-kube-api-access-krmjc\") pod \"route-controller-manager-6576b87f9c-xb7gk\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171803 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-service-ca\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171821 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4d69e866-000e-4748-9953-4d5f007f2669-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rnhfh\" (UID: \"4d69e866-000e-4748-9953-4d5f007f2669\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171839 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171857 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f25c1d0d-db4f-415b-a7ea-4d1b98e84819-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwk9c\" (UID: \"f25c1d0d-db4f-415b-a7ea-4d1b98e84819\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171876 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f893e28e-fb92-45b2-8c43-05ff99bf7d32-trusted-ca\") pod \"ingress-operator-5b745b69d9-gnrs8\" (UID: \"f893e28e-fb92-45b2-8c43-05ff99bf7d32\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171893 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e623ba7e-e936-4ab0-a825-6139b2b33fe4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kb5rn\" (UID: \"e623ba7e-e936-4ab0-a825-6139b2b33fe4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171909 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e60de411-4929-4a3a-a5ed-1071da7ed16a-config\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171926 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7dtd\" (UniqueName: \"kubernetes.io/projected/97042267-61ca-4864-a50a-60587abc7619-kube-api-access-j7dtd\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.172250 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-client-ca\") pod \"route-controller-manager-6576b87f9c-xb7gk\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.172634 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-client-ca\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.172889 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f25c1d0d-db4f-415b-a7ea-4d1b98e84819-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwk9c\" (UID: \"f25c1d0d-db4f-415b-a7ea-4d1b98e84819\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.173108 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e60de411-4929-4a3a-a5ed-1071da7ed16a-etcd-service-ca\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.173151 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.173500 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-trusted-ca-bundle\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.173647 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c225ab90-51ee-400f-8902-774ad6270322-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-678qm\" (UID: \"c225ab90-51ee-400f-8902-774ad6270322\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.173697 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e60de411-4929-4a3a-a5ed-1071da7ed16a-config\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.174295 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-service-ca\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.174305 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-console-config\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.174452 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4d69e866-000e-4748-9953-4d5f007f2669-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rnhfh\" (UID: \"4d69e866-000e-4748-9953-4d5f007f2669\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.175218 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-config\") pod \"route-controller-manager-6576b87f9c-xb7gk\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.171944 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06f03ec3-796d-442f-8adc-93aa22466426-config\") pod \"kube-controller-manager-operator-78b949d7b-fszts\" (UID: \"06f03ec3-796d-442f-8adc-93aa22466426\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.176917 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/97042267-61ca-4864-a50a-60587abc7619-console-serving-cert\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.176971 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-serving-cert\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177102 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e858c9d3-3f1c-47db-a0c8-f0efac1ed114-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-r7wlb\" (UID: \"e858c9d3-3f1c-47db-a0c8-f0efac1ed114\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r7wlb" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177198 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9crtj\" (UniqueName: \"kubernetes.io/projected/c225ab90-51ee-400f-8902-774ad6270322-kube-api-access-9crtj\") pod \"cluster-image-registry-operator-dc59b4c8b-678qm\" (UID: \"c225ab90-51ee-400f-8902-774ad6270322\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177316 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e60de411-4929-4a3a-a5ed-1071da7ed16a-etcd-client\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177331 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f893e28e-fb92-45b2-8c43-05ff99bf7d32-metrics-tls\") pod \"ingress-operator-5b745b69d9-gnrs8\" (UID: \"f893e28e-fb92-45b2-8c43-05ff99bf7d32\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177399 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e60de411-4929-4a3a-a5ed-1071da7ed16a-etcd-ca\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177459 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4d69e866-000e-4748-9953-4d5f007f2669-proxy-tls\") pod \"machine-config-controller-84d6567774-rnhfh\" (UID: \"4d69e866-000e-4748-9953-4d5f007f2669\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177479 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e623ba7e-e936-4ab0-a825-6139b2b33fe4-config\") pod \"kube-apiserver-operator-766d6c64bb-kb5rn\" (UID: \"e623ba7e-e936-4ab0-a825-6139b2b33fe4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177528 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-serving-cert\") pod \"route-controller-manager-6576b87f9c-xb7gk\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177566 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/97042267-61ca-4864-a50a-60587abc7619-console-oauth-config\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177620 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f893e28e-fb92-45b2-8c43-05ff99bf7d32-bound-sa-token\") pod \"ingress-operator-5b745b69d9-gnrs8\" (UID: \"f893e28e-fb92-45b2-8c43-05ff99bf7d32\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177646 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fktpj\" (UniqueName: \"kubernetes.io/projected/e60de411-4929-4a3a-a5ed-1071da7ed16a-kube-api-access-fktpj\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177689 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c225ab90-51ee-400f-8902-774ad6270322-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-678qm\" (UID: \"c225ab90-51ee-400f-8902-774ad6270322\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177720 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13187e9e-9b78-41ce-a2ff-7837dfd8eca6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fx54r\" (UID: \"13187e9e-9b78-41ce-a2ff-7837dfd8eca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177763 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-oauth-serving-cert\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177778 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c225ab90-51ee-400f-8902-774ad6270322-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-678qm\" (UID: \"c225ab90-51ee-400f-8902-774ad6270322\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.177802 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f25c1d0d-db4f-415b-a7ea-4d1b98e84819-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwk9c\" (UID: \"f25c1d0d-db4f-415b-a7ea-4d1b98e84819\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.178746 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx9d2\" (UniqueName: \"kubernetes.io/projected/13187e9e-9b78-41ce-a2ff-7837dfd8eca6-kube-api-access-xx9d2\") pod \"kube-storage-version-migrator-operator-b67b599dd-fx54r\" (UID: \"13187e9e-9b78-41ce-a2ff-7837dfd8eca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.178802 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06f03ec3-796d-442f-8adc-93aa22466426-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fszts\" (UID: \"06f03ec3-796d-442f-8adc-93aa22466426\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.178829 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-config\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.178890 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06f03ec3-796d-442f-8adc-93aa22466426-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fszts\" (UID: \"06f03ec3-796d-442f-8adc-93aa22466426\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.178917 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngjxb\" (UniqueName: \"kubernetes.io/projected/e858c9d3-3f1c-47db-a0c8-f0efac1ed114-kube-api-access-ngjxb\") pod \"multus-admission-controller-857f4d67dd-r7wlb\" (UID: \"e858c9d3-3f1c-47db-a0c8-f0efac1ed114\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r7wlb" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.178990 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e60de411-4929-4a3a-a5ed-1071da7ed16a-serving-cert\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.179008 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqnvq\" (UniqueName: \"kubernetes.io/projected/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-kube-api-access-hqnvq\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.179051 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b5f5\" (UniqueName: \"kubernetes.io/projected/4d69e866-000e-4748-9953-4d5f007f2669-kube-api-access-4b5f5\") pod \"machine-config-controller-84d6567774-rnhfh\" (UID: \"4d69e866-000e-4748-9953-4d5f007f2669\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.179556 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-oauth-serving-cert\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.180225 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-config\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.180822 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-serving-cert\") pod \"route-controller-manager-6576b87f9c-xb7gk\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.180882 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f25c1d0d-db4f-415b-a7ea-4d1b98e84819-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwk9c\" (UID: \"f25c1d0d-db4f-415b-a7ea-4d1b98e84819\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.181785 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c225ab90-51ee-400f-8902-774ad6270322-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-678qm\" (UID: \"c225ab90-51ee-400f-8902-774ad6270322\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.183102 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e60de411-4929-4a3a-a5ed-1071da7ed16a-serving-cert\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.184722 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/97042267-61ca-4864-a50a-60587abc7619-console-oauth-config\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.189048 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.207752 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.218808 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e60de411-4929-4a3a-a5ed-1071da7ed16a-etcd-ca\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.227728 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.248063 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.268140 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.272352 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13187e9e-9b78-41ce-a2ff-7837dfd8eca6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fx54r\" (UID: \"13187e9e-9b78-41ce-a2ff-7837dfd8eca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.288559 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.308422 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.327726 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.334665 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13187e9e-9b78-41ce-a2ff-7837dfd8eca6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fx54r\" (UID: \"13187e9e-9b78-41ce-a2ff-7837dfd8eca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.349059 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.362788 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f893e28e-fb92-45b2-8c43-05ff99bf7d32-metrics-tls\") pod \"ingress-operator-5b745b69d9-gnrs8\" (UID: \"f893e28e-fb92-45b2-8c43-05ff99bf7d32\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.368839 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.395879 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.405360 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f893e28e-fb92-45b2-8c43-05ff99bf7d32-trusted-ca\") pod \"ingress-operator-5b745b69d9-gnrs8\" (UID: \"f893e28e-fb92-45b2-8c43-05ff99bf7d32\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.410168 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.429818 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.437572 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e623ba7e-e936-4ab0-a825-6139b2b33fe4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kb5rn\" (UID: \"e623ba7e-e936-4ab0-a825-6139b2b33fe4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.449747 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.468335 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.489261 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.498991 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e623ba7e-e936-4ab0-a825-6139b2b33fe4-config\") pod \"kube-apiserver-operator-766d6c64bb-kb5rn\" (UID: \"e623ba7e-e936-4ab0-a825-6139b2b33fe4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.510616 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.529840 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.549633 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.569101 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.589130 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.602409 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4d69e866-000e-4748-9953-4d5f007f2669-proxy-tls\") pod \"machine-config-controller-84d6567774-rnhfh\" (UID: \"4d69e866-000e-4748-9953-4d5f007f2669\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.608322 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.629058 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.648341 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.669646 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.681352 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e858c9d3-3f1c-47db-a0c8-f0efac1ed114-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-r7wlb\" (UID: \"e858c9d3-3f1c-47db-a0c8-f0efac1ed114\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r7wlb" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.688270 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.708776 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.714836 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06f03ec3-796d-442f-8adc-93aa22466426-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fszts\" (UID: \"06f03ec3-796d-442f-8adc-93aa22466426\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.728157 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.734496 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06f03ec3-796d-442f-8adc-93aa22466426-config\") pod \"kube-controller-manager-operator-78b949d7b-fszts\" (UID: \"06f03ec3-796d-442f-8adc-93aa22466426\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.768982 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.788645 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.809648 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.829288 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.849602 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.869020 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.889463 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.909299 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.928862 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.948447 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.966945 4806 request.go:700] Waited for 1.000774415s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dolm-operator-serviceaccount-dockercfg-rq7zk&limit=500&resourceVersion=0 Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.968992 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 27 10:23:55 crc kubenswrapper[4806]: I1127 10:23:55.989627 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.009082 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.028114 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.049593 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.069894 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.097541 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.108851 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.128667 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.149012 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.169033 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.189035 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.209390 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.229321 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.249340 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.269925 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.289466 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.309332 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.329759 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.349412 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.369569 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.389279 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.409278 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.428168 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.449821 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.469878 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.490134 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.509358 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.532145 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.550756 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.569407 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.590004 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.609348 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.629522 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.648992 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.668222 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.690542 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.710464 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.729103 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.749436 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.798332 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzkwd\" (UniqueName: \"kubernetes.io/projected/47d1375e-f8bc-4406-8692-fe9aecb0e394-kube-api-access-mzkwd\") pod \"apiserver-7bbb656c7d-bsqsc\" (UID: \"47d1375e-f8bc-4406-8692-fe9aecb0e394\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.820046 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tphxc\" (UniqueName: \"kubernetes.io/projected/52dd6d6d-3a15-4467-9d82-37fd2fa5794b-kube-api-access-tphxc\") pod \"downloads-7954f5f757-9ppfs\" (UID: \"52dd6d6d-3a15-4467-9d82-37fd2fa5794b\") " pod="openshift-console/downloads-7954f5f757-9ppfs" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.840797 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x49vx\" (UniqueName: \"kubernetes.io/projected/d26adb0c-6bf0-49f1-806b-06d96e333fc4-kube-api-access-x49vx\") pod \"machine-approver-56656f9798-sqqv6\" (UID: \"d26adb0c-6bf0-49f1-806b-06d96e333fc4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.859131 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjm5h\" (UniqueName: \"kubernetes.io/projected/a52b3a70-665d-49e5-a842-14ade4dcff1f-kube-api-access-jjm5h\") pod \"apiserver-76f77b778f-8jh8g\" (UID: \"a52b3a70-665d-49e5-a842-14ade4dcff1f\") " pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.870548 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcqvp\" (UniqueName: \"kubernetes.io/projected/52480926-8957-4c79-a384-3e2d27f349af-kube-api-access-wcqvp\") pod \"openshift-apiserver-operator-796bbdcf4f-8gdsx\" (UID: \"52480926-8957-4c79-a384-3e2d27f349af\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.888196 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qntsv\" (UniqueName: \"kubernetes.io/projected/6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe-kube-api-access-qntsv\") pod \"openshift-controller-manager-operator-756b6f6bc6-dqjqc\" (UID: \"6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.904670 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82hcz\" (UniqueName: \"kubernetes.io/projected/325e82c6-eacf-4180-bc5f-9334536ebe16-kube-api-access-82hcz\") pod \"machine-api-operator-5694c8668f-zd4hv\" (UID: \"325e82c6-eacf-4180-bc5f-9334536ebe16\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.927744 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx8kd\" (UniqueName: \"kubernetes.io/projected/65b0faa1-6054-43a4-ab15-e59791790337-kube-api-access-hx8kd\") pod \"openshift-config-operator-7777fb866f-pppcb\" (UID: \"65b0faa1-6054-43a4-ab15-e59791790337\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.929425 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.937645 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.947733 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqf8w\" (UniqueName: \"kubernetes.io/projected/a3ce016c-a670-46a3-9254-03fea2c6568b-kube-api-access-sqf8w\") pod \"cluster-samples-operator-665b6dd947-qgxx9\" (UID: \"a3ce016c-a670-46a3-9254-03fea2c6568b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.970020 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7qs6\" (UniqueName: \"kubernetes.io/projected/0639fe21-6b99-4fad-be7b-6bd998c3472c-kube-api-access-n7qs6\") pod \"oauth-openshift-558db77b4-l956q\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.974006 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.986519 4806 request.go:700] Waited for 1.915288047s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console-operator/serviceaccounts/console-operator/token Nov 27 10:23:56 crc kubenswrapper[4806]: I1127 10:23:56.990000 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qxs4\" (UniqueName: \"kubernetes.io/projected/93b844fb-455e-4a0a-9d75-52f1278df25c-kube-api-access-5qxs4\") pod \"authentication-operator-69f744f599-z4ndr\" (UID: \"93b844fb-455e-4a0a-9d75-52f1278df25c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.007150 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hbnm\" (UniqueName: \"kubernetes.io/projected/d52f4ad9-009f-41cf-a779-2a6515a2fbb4-kube-api-access-7hbnm\") pod \"console-operator-58897d9998-42h4s\" (UID: \"d52f4ad9-009f-41cf-a779-2a6515a2fbb4\") " pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.007175 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.031429 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.036000 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.047088 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9ppfs" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.050503 4806 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.060073 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.068867 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.069442 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.084402 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.100341 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.111113 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqslv\" (UniqueName: \"kubernetes.io/projected/f893e28e-fb92-45b2-8c43-05ff99bf7d32-kube-api-access-kqslv\") pod \"ingress-operator-5b745b69d9-gnrs8\" (UID: \"f893e28e-fb92-45b2-8c43-05ff99bf7d32\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.130430 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e623ba7e-e936-4ab0-a825-6139b2b33fe4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kb5rn\" (UID: \"e623ba7e-e936-4ab0-a825-6139b2b33fe4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.143574 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f25c1d0d-db4f-415b-a7ea-4d1b98e84819-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwk9c\" (UID: \"f25c1d0d-db4f-415b-a7ea-4d1b98e84819\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.150297 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.169384 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7dtd\" (UniqueName: \"kubernetes.io/projected/97042267-61ca-4864-a50a-60587abc7619-kube-api-access-j7dtd\") pod \"console-f9d7485db-6fqzk\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.196222 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krmjc\" (UniqueName: \"kubernetes.io/projected/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-kube-api-access-krmjc\") pod \"route-controller-manager-6576b87f9c-xb7gk\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.207752 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.210958 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9crtj\" (UniqueName: \"kubernetes.io/projected/c225ab90-51ee-400f-8902-774ad6270322-kube-api-access-9crtj\") pod \"cluster-image-registry-operator-dc59b4c8b-678qm\" (UID: \"c225ab90-51ee-400f-8902-774ad6270322\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.213637 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8jh8g"] Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.217610 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.220377 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.230792 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c225ab90-51ee-400f-8902-774ad6270322-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-678qm\" (UID: \"c225ab90-51ee-400f-8902-774ad6270322\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.257277 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f893e28e-fb92-45b2-8c43-05ff99bf7d32-bound-sa-token\") pod \"ingress-operator-5b745b69d9-gnrs8\" (UID: \"f893e28e-fb92-45b2-8c43-05ff99bf7d32\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.271408 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fktpj\" (UniqueName: \"kubernetes.io/projected/e60de411-4929-4a3a-a5ed-1071da7ed16a-kube-api-access-fktpj\") pod \"etcd-operator-b45778765-8jvb8\" (UID: \"e60de411-4929-4a3a-a5ed-1071da7ed16a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.280935 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-pppcb"] Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.289426 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx9d2\" (UniqueName: \"kubernetes.io/projected/13187e9e-9b78-41ce-a2ff-7837dfd8eca6-kube-api-access-xx9d2\") pod \"kube-storage-version-migrator-operator-b67b599dd-fx54r\" (UID: \"13187e9e-9b78-41ce-a2ff-7837dfd8eca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.309700 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b5f5\" (UniqueName: \"kubernetes.io/projected/4d69e866-000e-4748-9953-4d5f007f2669-kube-api-access-4b5f5\") pod \"machine-config-controller-84d6567774-rnhfh\" (UID: \"4d69e866-000e-4748-9953-4d5f007f2669\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.324652 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06f03ec3-796d-442f-8adc-93aa22466426-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fszts\" (UID: \"06f03ec3-796d-442f-8adc-93aa22466426\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.347135 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx"] Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.353156 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngjxb\" (UniqueName: \"kubernetes.io/projected/e858c9d3-3f1c-47db-a0c8-f0efac1ed114-kube-api-access-ngjxb\") pod \"multus-admission-controller-857f4d67dd-r7wlb\" (UID: \"e858c9d3-3f1c-47db-a0c8-f0efac1ed114\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r7wlb" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.371592 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqnvq\" (UniqueName: \"kubernetes.io/projected/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-kube-api-access-hqnvq\") pod \"controller-manager-879f6c89f-2szrd\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.409887 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.421465 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2a6f027-898d-4a60-aa19-00b6f54d2aac-trusted-ca\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.421525 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f9ef04e5-e987-4457-bd8b-d1bc01cb67ac-metrics-tls\") pod \"dns-operator-744455d44c-lzwl4\" (UID: \"f9ef04e5-e987-4457-bd8b-d1bc01cb67ac\") " pod="openshift-dns-operator/dns-operator-744455d44c-lzwl4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.421561 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlz4q\" (UniqueName: \"kubernetes.io/projected/f9ef04e5-e987-4457-bd8b-d1bc01cb67ac-kube-api-access-xlz4q\") pod \"dns-operator-744455d44c-lzwl4\" (UID: \"f9ef04e5-e987-4457-bd8b-d1bc01cb67ac\") " pod="openshift-dns-operator/dns-operator-744455d44c-lzwl4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.421612 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.421635 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2a6f027-898d-4a60-aa19-00b6f54d2aac-registry-certificates\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.421671 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-bound-sa-token\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.421692 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2a6f027-898d-4a60-aa19-00b6f54d2aac-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.421711 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2a6f027-898d-4a60-aa19-00b6f54d2aac-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.421736 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrlcf\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-kube-api-access-lrlcf\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.421788 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-registry-tls\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: E1127 10:23:57.422267 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:23:57.922248878 +0000 UTC m=+142.508839642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.423359 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.424948 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.425716 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9ppfs"] Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.435315 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.500065 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.506169 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.512774 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532128 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532512 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea7b78e0-1c47-45a1-b93c-bd7eba502143-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-sg6x8\" (UID: \"ea7b78e0-1c47-45a1-b93c-bd7eba502143\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532583 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvj2s\" (UniqueName: \"kubernetes.io/projected/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-kube-api-access-hvj2s\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532602 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9lgw\" (UniqueName: \"kubernetes.io/projected/cc69668d-b5d6-452a-99bf-b9a14c90a65d-kube-api-access-g9lgw\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532678 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2a6f027-898d-4a60-aa19-00b6f54d2aac-trusted-ca\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532698 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-plugins-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532732 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc69668d-b5d6-452a-99bf-b9a14c90a65d-metrics-certs\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532762 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/16a0adb3-9386-417e-8a34-bdf429462e2b-apiservice-cert\") pod \"packageserver-d55dfcdfc-wsjh5\" (UID: \"16a0adb3-9386-417e-8a34-bdf429462e2b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532789 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2b8p\" (UniqueName: \"kubernetes.io/projected/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-kube-api-access-d2b8p\") pod \"marketplace-operator-79b997595-7t7px\" (UID: \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\") " pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532832 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsddn\" (UniqueName: \"kubernetes.io/projected/777fcbda-0e10-4a62-aa45-fa8670bae076-kube-api-access-bsddn\") pod \"service-ca-operator-777779d784-nflhv\" (UID: \"777fcbda-0e10-4a62-aa45-fa8670bae076\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532873 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnbph\" (UniqueName: \"kubernetes.io/projected/41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693-kube-api-access-bnbph\") pod \"machine-config-server-gwfw8\" (UID: \"41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693\") " pod="openshift-machine-config-operator/machine-config-server-gwfw8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532903 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f9ef04e5-e987-4457-bd8b-d1bc01cb67ac-metrics-tls\") pod \"dns-operator-744455d44c-lzwl4\" (UID: \"f9ef04e5-e987-4457-bd8b-d1bc01cb67ac\") " pod="openshift-dns-operator/dns-operator-744455d44c-lzwl4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532931 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgcdp\" (UniqueName: \"kubernetes.io/projected/9b0fd0cb-13fa-4f87-bfc3-defe908884d4-kube-api-access-wgcdp\") pod \"service-ca-9c57cc56f-28rbb\" (UID: \"9b0fd0cb-13fa-4f87-bfc3-defe908884d4\") " pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.532995 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f6fc593a-b4b9-46ee-b12f-9ac4732dc69e-metrics-tls\") pod \"dns-default-zbhcm\" (UID: \"f6fc593a-b4b9-46ee-b12f-9ac4732dc69e\") " pod="openshift-dns/dns-default-zbhcm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.533033 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7t7px\" (UID: \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\") " pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:23:57 crc kubenswrapper[4806]: E1127 10:23:57.533085 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:23:58.033065009 +0000 UTC m=+142.619655773 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.534248 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2a6f027-898d-4a60-aa19-00b6f54d2aac-trusted-ca\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.534540 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.534552 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3fb2578f-5e34-45b4-bcae-4798bfba073a-profile-collector-cert\") pod \"catalog-operator-68c6474976-xvtnq\" (UID: \"3fb2578f-5e34-45b4-bcae-4798bfba073a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.534584 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlz4q\" (UniqueName: \"kubernetes.io/projected/f9ef04e5-e987-4457-bd8b-d1bc01cb67ac-kube-api-access-xlz4q\") pod \"dns-operator-744455d44c-lzwl4\" (UID: \"f9ef04e5-e987-4457-bd8b-d1bc01cb67ac\") " pod="openshift-dns-operator/dns-operator-744455d44c-lzwl4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.534672 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnjm4\" (UniqueName: \"kubernetes.io/projected/3c890acc-37e4-4838-92c5-ff9a72738519-kube-api-access-cnjm4\") pod \"migrator-59844c95c7-62tks\" (UID: \"3c890acc-37e4-4838-92c5-ff9a72738519\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-62tks" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.535727 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-mountpoint-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.535766 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvg48\" (UniqueName: \"kubernetes.io/projected/67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06-kube-api-access-gvg48\") pod \"machine-config-operator-74547568cd-prj6n\" (UID: \"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.535826 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nftbv\" (UniqueName: \"kubernetes.io/projected/a053f34f-9565-4d7c-88bc-2edd1c08cb69-kube-api-access-nftbv\") pod \"ingress-canary-57tr4\" (UID: \"a053f34f-9565-4d7c-88bc-2edd1c08cb69\") " pod="openshift-ingress-canary/ingress-canary-57tr4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.535843 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6tbp\" (UniqueName: \"kubernetes.io/projected/42e226ef-f2d6-4a0a-8d35-f9a095322a81-kube-api-access-p6tbp\") pod \"olm-operator-6b444d44fb-pxg4x\" (UID: \"42e226ef-f2d6-4a0a-8d35-f9a095322a81\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.535881 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-csi-data-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.535895 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/16a0adb3-9386-417e-8a34-bdf429462e2b-tmpfs\") pod \"packageserver-d55dfcdfc-wsjh5\" (UID: \"16a0adb3-9386-417e-8a34-bdf429462e2b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.535912 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cc69668d-b5d6-452a-99bf-b9a14c90a65d-default-certificate\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.535978 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbs8q\" (UniqueName: \"kubernetes.io/projected/7308dfe6-f141-4f20-a966-075d17882333-kube-api-access-rbs8q\") pod \"collect-profiles-29403975-mz4zk\" (UID: \"7308dfe6-f141-4f20-a966-075d17882333\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.536049 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a053f34f-9565-4d7c-88bc-2edd1c08cb69-cert\") pod \"ingress-canary-57tr4\" (UID: \"a053f34f-9565-4d7c-88bc-2edd1c08cb69\") " pod="openshift-ingress-canary/ingress-canary-57tr4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.536067 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06-images\") pod \"machine-config-operator-74547568cd-prj6n\" (UID: \"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.537418 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.537464 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693-certs\") pod \"machine-config-server-gwfw8\" (UID: \"41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693\") " pod="openshift-machine-config-operator/machine-config-server-gwfw8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.537600 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7zwh\" (UniqueName: \"kubernetes.io/projected/16a0adb3-9386-417e-8a34-bdf429462e2b-kube-api-access-w7zwh\") pod \"packageserver-d55dfcdfc-wsjh5\" (UID: \"16a0adb3-9386-417e-8a34-bdf429462e2b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.537649 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4fdd\" (UniqueName: \"kubernetes.io/projected/adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce-kube-api-access-l4fdd\") pod \"control-plane-machine-set-operator-78cbb6b69f-5qtlg\" (UID: \"adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg" Nov 27 10:23:57 crc kubenswrapper[4806]: E1127 10:23:57.537694 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:23:58.037678917 +0000 UTC m=+142.624269681 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.537730 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/777fcbda-0e10-4a62-aa45-fa8670bae076-serving-cert\") pod \"service-ca-operator-777779d784-nflhv\" (UID: \"777fcbda-0e10-4a62-aa45-fa8670bae076\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.537790 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2a6f027-898d-4a60-aa19-00b6f54d2aac-registry-certificates\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.537819 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3fb2578f-5e34-45b4-bcae-4798bfba073a-srv-cert\") pod \"catalog-operator-68c6474976-xvtnq\" (UID: \"3fb2578f-5e34-45b4-bcae-4798bfba073a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.538671 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2a6f027-898d-4a60-aa19-00b6f54d2aac-registry-certificates\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.538723 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-bound-sa-token\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539399 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2a6f027-898d-4a60-aa19-00b6f54d2aac-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539453 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2a6f027-898d-4a60-aa19-00b6f54d2aac-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539471 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06-auth-proxy-config\") pod \"machine-config-operator-74547568cd-prj6n\" (UID: \"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539530 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-registration-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539546 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/16a0adb3-9386-417e-8a34-bdf429462e2b-webhook-cert\") pod \"packageserver-d55dfcdfc-wsjh5\" (UID: \"16a0adb3-9386-417e-8a34-bdf429462e2b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539562 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kpvd\" (UniqueName: \"kubernetes.io/projected/f6fc593a-b4b9-46ee-b12f-9ac4732dc69e-kube-api-access-9kpvd\") pod \"dns-default-zbhcm\" (UID: \"f6fc593a-b4b9-46ee-b12f-9ac4732dc69e\") " pod="openshift-dns/dns-default-zbhcm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539591 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrlcf\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-kube-api-access-lrlcf\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539607 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9b0fd0cb-13fa-4f87-bfc3-defe908884d4-signing-cabundle\") pod \"service-ca-9c57cc56f-28rbb\" (UID: \"9b0fd0cb-13fa-4f87-bfc3-defe908884d4\") " pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539698 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5qtlg\" (UID: \"adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539744 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6fc593a-b4b9-46ee-b12f-9ac4732dc69e-config-volume\") pod \"dns-default-zbhcm\" (UID: \"f6fc593a-b4b9-46ee-b12f-9ac4732dc69e\") " pod="openshift-dns/dns-default-zbhcm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539763 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb2fx\" (UniqueName: \"kubernetes.io/projected/3fb2578f-5e34-45b4-bcae-4798bfba073a-kube-api-access-zb2fx\") pod \"catalog-operator-68c6474976-xvtnq\" (UID: \"3fb2578f-5e34-45b4-bcae-4798bfba073a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539781 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/777fcbda-0e10-4a62-aa45-fa8670bae076-config\") pod \"service-ca-operator-777779d784-nflhv\" (UID: \"777fcbda-0e10-4a62-aa45-fa8670bae076\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539795 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7308dfe6-f141-4f20-a966-075d17882333-secret-volume\") pod \"collect-profiles-29403975-mz4zk\" (UID: \"7308dfe6-f141-4f20-a966-075d17882333\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539813 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7t7px\" (UID: \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\") " pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539827 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-socket-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.539899 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9b0fd0cb-13fa-4f87-bfc3-defe908884d4-signing-key\") pod \"service-ca-9c57cc56f-28rbb\" (UID: \"9b0fd0cb-13fa-4f87-bfc3-defe908884d4\") " pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.540014 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cc69668d-b5d6-452a-99bf-b9a14c90a65d-stats-auth\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.540049 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693-node-bootstrap-token\") pod \"machine-config-server-gwfw8\" (UID: \"41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693\") " pod="openshift-machine-config-operator/machine-config-server-gwfw8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.540220 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnn7k\" (UniqueName: \"kubernetes.io/projected/ea7b78e0-1c47-45a1-b93c-bd7eba502143-kube-api-access-nnn7k\") pod \"package-server-manager-789f6589d5-sg6x8\" (UID: \"ea7b78e0-1c47-45a1-b93c-bd7eba502143\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.540266 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06-proxy-tls\") pod \"machine-config-operator-74547568cd-prj6n\" (UID: \"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.540295 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/42e226ef-f2d6-4a0a-8d35-f9a095322a81-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pxg4x\" (UID: \"42e226ef-f2d6-4a0a-8d35-f9a095322a81\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.540311 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc69668d-b5d6-452a-99bf-b9a14c90a65d-service-ca-bundle\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.540328 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7308dfe6-f141-4f20-a966-075d17882333-config-volume\") pod \"collect-profiles-29403975-mz4zk\" (UID: \"7308dfe6-f141-4f20-a966-075d17882333\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.540344 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/42e226ef-f2d6-4a0a-8d35-f9a095322a81-srv-cert\") pod \"olm-operator-6b444d44fb-pxg4x\" (UID: \"42e226ef-f2d6-4a0a-8d35-f9a095322a81\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.540416 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-registry-tls\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.542070 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2a6f027-898d-4a60-aa19-00b6f54d2aac-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.545291 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-r7wlb" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.549373 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.549924 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2a6f027-898d-4a60-aa19-00b6f54d2aac-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.551881 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-registry-tls\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.552125 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f9ef04e5-e987-4457-bd8b-d1bc01cb67ac-metrics-tls\") pod \"dns-operator-744455d44c-lzwl4\" (UID: \"f9ef04e5-e987-4457-bd8b-d1bc01cb67ac\") " pod="openshift-dns-operator/dns-operator-744455d44c-lzwl4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.570983 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlz4q\" (UniqueName: \"kubernetes.io/projected/f9ef04e5-e987-4457-bd8b-d1bc01cb67ac-kube-api-access-xlz4q\") pod \"dns-operator-744455d44c-lzwl4\" (UID: \"f9ef04e5-e987-4457-bd8b-d1bc01cb67ac\") " pod="openshift-dns-operator/dns-operator-744455d44c-lzwl4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.614683 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-bound-sa-token\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.627851 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrlcf\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-kube-api-access-lrlcf\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642220 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642484 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cc69668d-b5d6-452a-99bf-b9a14c90a65d-stats-auth\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642513 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693-node-bootstrap-token\") pod \"machine-config-server-gwfw8\" (UID: \"41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693\") " pod="openshift-machine-config-operator/machine-config-server-gwfw8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642532 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnn7k\" (UniqueName: \"kubernetes.io/projected/ea7b78e0-1c47-45a1-b93c-bd7eba502143-kube-api-access-nnn7k\") pod \"package-server-manager-789f6589d5-sg6x8\" (UID: \"ea7b78e0-1c47-45a1-b93c-bd7eba502143\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642555 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06-proxy-tls\") pod \"machine-config-operator-74547568cd-prj6n\" (UID: \"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642572 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7308dfe6-f141-4f20-a966-075d17882333-config-volume\") pod \"collect-profiles-29403975-mz4zk\" (UID: \"7308dfe6-f141-4f20-a966-075d17882333\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642588 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/42e226ef-f2d6-4a0a-8d35-f9a095322a81-srv-cert\") pod \"olm-operator-6b444d44fb-pxg4x\" (UID: \"42e226ef-f2d6-4a0a-8d35-f9a095322a81\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642602 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/42e226ef-f2d6-4a0a-8d35-f9a095322a81-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pxg4x\" (UID: \"42e226ef-f2d6-4a0a-8d35-f9a095322a81\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642619 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc69668d-b5d6-452a-99bf-b9a14c90a65d-service-ca-bundle\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642645 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea7b78e0-1c47-45a1-b93c-bd7eba502143-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-sg6x8\" (UID: \"ea7b78e0-1c47-45a1-b93c-bd7eba502143\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642667 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvj2s\" (UniqueName: \"kubernetes.io/projected/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-kube-api-access-hvj2s\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642690 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9lgw\" (UniqueName: \"kubernetes.io/projected/cc69668d-b5d6-452a-99bf-b9a14c90a65d-kube-api-access-g9lgw\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642713 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-plugins-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642727 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc69668d-b5d6-452a-99bf-b9a14c90a65d-metrics-certs\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642747 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/16a0adb3-9386-417e-8a34-bdf429462e2b-apiservice-cert\") pod \"packageserver-d55dfcdfc-wsjh5\" (UID: \"16a0adb3-9386-417e-8a34-bdf429462e2b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642766 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2b8p\" (UniqueName: \"kubernetes.io/projected/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-kube-api-access-d2b8p\") pod \"marketplace-operator-79b997595-7t7px\" (UID: \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\") " pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642785 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsddn\" (UniqueName: \"kubernetes.io/projected/777fcbda-0e10-4a62-aa45-fa8670bae076-kube-api-access-bsddn\") pod \"service-ca-operator-777779d784-nflhv\" (UID: \"777fcbda-0e10-4a62-aa45-fa8670bae076\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642803 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnbph\" (UniqueName: \"kubernetes.io/projected/41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693-kube-api-access-bnbph\") pod \"machine-config-server-gwfw8\" (UID: \"41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693\") " pod="openshift-machine-config-operator/machine-config-server-gwfw8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642822 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgcdp\" (UniqueName: \"kubernetes.io/projected/9b0fd0cb-13fa-4f87-bfc3-defe908884d4-kube-api-access-wgcdp\") pod \"service-ca-9c57cc56f-28rbb\" (UID: \"9b0fd0cb-13fa-4f87-bfc3-defe908884d4\") " pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642842 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f6fc593a-b4b9-46ee-b12f-9ac4732dc69e-metrics-tls\") pod \"dns-default-zbhcm\" (UID: \"f6fc593a-b4b9-46ee-b12f-9ac4732dc69e\") " pod="openshift-dns/dns-default-zbhcm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642861 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7t7px\" (UID: \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\") " pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642890 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3fb2578f-5e34-45b4-bcae-4798bfba073a-profile-collector-cert\") pod \"catalog-operator-68c6474976-xvtnq\" (UID: \"3fb2578f-5e34-45b4-bcae-4798bfba073a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642907 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnjm4\" (UniqueName: \"kubernetes.io/projected/3c890acc-37e4-4838-92c5-ff9a72738519-kube-api-access-cnjm4\") pod \"migrator-59844c95c7-62tks\" (UID: \"3c890acc-37e4-4838-92c5-ff9a72738519\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-62tks" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642924 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-mountpoint-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642939 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvg48\" (UniqueName: \"kubernetes.io/projected/67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06-kube-api-access-gvg48\") pod \"machine-config-operator-74547568cd-prj6n\" (UID: \"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642956 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nftbv\" (UniqueName: \"kubernetes.io/projected/a053f34f-9565-4d7c-88bc-2edd1c08cb69-kube-api-access-nftbv\") pod \"ingress-canary-57tr4\" (UID: \"a053f34f-9565-4d7c-88bc-2edd1c08cb69\") " pod="openshift-ingress-canary/ingress-canary-57tr4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642978 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6tbp\" (UniqueName: \"kubernetes.io/projected/42e226ef-f2d6-4a0a-8d35-f9a095322a81-kube-api-access-p6tbp\") pod \"olm-operator-6b444d44fb-pxg4x\" (UID: \"42e226ef-f2d6-4a0a-8d35-f9a095322a81\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.642999 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/16a0adb3-9386-417e-8a34-bdf429462e2b-tmpfs\") pod \"packageserver-d55dfcdfc-wsjh5\" (UID: \"16a0adb3-9386-417e-8a34-bdf429462e2b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643014 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cc69668d-b5d6-452a-99bf-b9a14c90a65d-default-certificate\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643029 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-csi-data-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643045 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbs8q\" (UniqueName: \"kubernetes.io/projected/7308dfe6-f141-4f20-a966-075d17882333-kube-api-access-rbs8q\") pod \"collect-profiles-29403975-mz4zk\" (UID: \"7308dfe6-f141-4f20-a966-075d17882333\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643075 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06-images\") pod \"machine-config-operator-74547568cd-prj6n\" (UID: \"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643089 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a053f34f-9565-4d7c-88bc-2edd1c08cb69-cert\") pod \"ingress-canary-57tr4\" (UID: \"a053f34f-9565-4d7c-88bc-2edd1c08cb69\") " pod="openshift-ingress-canary/ingress-canary-57tr4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643116 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693-certs\") pod \"machine-config-server-gwfw8\" (UID: \"41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693\") " pod="openshift-machine-config-operator/machine-config-server-gwfw8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643134 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7zwh\" (UniqueName: \"kubernetes.io/projected/16a0adb3-9386-417e-8a34-bdf429462e2b-kube-api-access-w7zwh\") pod \"packageserver-d55dfcdfc-wsjh5\" (UID: \"16a0adb3-9386-417e-8a34-bdf429462e2b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643149 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4fdd\" (UniqueName: \"kubernetes.io/projected/adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce-kube-api-access-l4fdd\") pod \"control-plane-machine-set-operator-78cbb6b69f-5qtlg\" (UID: \"adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643165 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3fb2578f-5e34-45b4-bcae-4798bfba073a-srv-cert\") pod \"catalog-operator-68c6474976-xvtnq\" (UID: \"3fb2578f-5e34-45b4-bcae-4798bfba073a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643180 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/777fcbda-0e10-4a62-aa45-fa8670bae076-serving-cert\") pod \"service-ca-operator-777779d784-nflhv\" (UID: \"777fcbda-0e10-4a62-aa45-fa8670bae076\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643197 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06-auth-proxy-config\") pod \"machine-config-operator-74547568cd-prj6n\" (UID: \"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643211 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/16a0adb3-9386-417e-8a34-bdf429462e2b-webhook-cert\") pod \"packageserver-d55dfcdfc-wsjh5\" (UID: \"16a0adb3-9386-417e-8a34-bdf429462e2b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643243 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kpvd\" (UniqueName: \"kubernetes.io/projected/f6fc593a-b4b9-46ee-b12f-9ac4732dc69e-kube-api-access-9kpvd\") pod \"dns-default-zbhcm\" (UID: \"f6fc593a-b4b9-46ee-b12f-9ac4732dc69e\") " pod="openshift-dns/dns-default-zbhcm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643260 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-registration-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643275 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9b0fd0cb-13fa-4f87-bfc3-defe908884d4-signing-cabundle\") pod \"service-ca-9c57cc56f-28rbb\" (UID: \"9b0fd0cb-13fa-4f87-bfc3-defe908884d4\") " pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643292 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5qtlg\" (UID: \"adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643309 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb2fx\" (UniqueName: \"kubernetes.io/projected/3fb2578f-5e34-45b4-bcae-4798bfba073a-kube-api-access-zb2fx\") pod \"catalog-operator-68c6474976-xvtnq\" (UID: \"3fb2578f-5e34-45b4-bcae-4798bfba073a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643324 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/777fcbda-0e10-4a62-aa45-fa8670bae076-config\") pod \"service-ca-operator-777779d784-nflhv\" (UID: \"777fcbda-0e10-4a62-aa45-fa8670bae076\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643338 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6fc593a-b4b9-46ee-b12f-9ac4732dc69e-config-volume\") pod \"dns-default-zbhcm\" (UID: \"f6fc593a-b4b9-46ee-b12f-9ac4732dc69e\") " pod="openshift-dns/dns-default-zbhcm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643355 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7308dfe6-f141-4f20-a966-075d17882333-secret-volume\") pod \"collect-profiles-29403975-mz4zk\" (UID: \"7308dfe6-f141-4f20-a966-075d17882333\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643369 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7t7px\" (UID: \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\") " pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643384 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-socket-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.643406 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9b0fd0cb-13fa-4f87-bfc3-defe908884d4-signing-key\") pod \"service-ca-9c57cc56f-28rbb\" (UID: \"9b0fd0cb-13fa-4f87-bfc3-defe908884d4\") " pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" Nov 27 10:23:57 crc kubenswrapper[4806]: E1127 10:23:57.643702 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:23:58.143687623 +0000 UTC m=+142.730278377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.651702 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06-auth-proxy-config\") pod \"machine-config-operator-74547568cd-prj6n\" (UID: \"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.657061 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/777fcbda-0e10-4a62-aa45-fa8670bae076-serving-cert\") pod \"service-ca-operator-777779d784-nflhv\" (UID: \"777fcbda-0e10-4a62-aa45-fa8670bae076\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.657425 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/16a0adb3-9386-417e-8a34-bdf429462e2b-apiservice-cert\") pod \"packageserver-d55dfcdfc-wsjh5\" (UID: \"16a0adb3-9386-417e-8a34-bdf429462e2b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.657577 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/16a0adb3-9386-417e-8a34-bdf429462e2b-tmpfs\") pod \"packageserver-d55dfcdfc-wsjh5\" (UID: \"16a0adb3-9386-417e-8a34-bdf429462e2b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.658167 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-registration-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.660109 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9b0fd0cb-13fa-4f87-bfc3-defe908884d4-signing-key\") pod \"service-ca-9c57cc56f-28rbb\" (UID: \"9b0fd0cb-13fa-4f87-bfc3-defe908884d4\") " pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.661096 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/16a0adb3-9386-417e-8a34-bdf429462e2b-webhook-cert\") pod \"packageserver-d55dfcdfc-wsjh5\" (UID: \"16a0adb3-9386-417e-8a34-bdf429462e2b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.661395 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc69668d-b5d6-452a-99bf-b9a14c90a65d-metrics-certs\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.661597 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9b0fd0cb-13fa-4f87-bfc3-defe908884d4-signing-cabundle\") pod \"service-ca-9c57cc56f-28rbb\" (UID: \"9b0fd0cb-13fa-4f87-bfc3-defe908884d4\") " pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.662879 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f6fc593a-b4b9-46ee-b12f-9ac4732dc69e-metrics-tls\") pod \"dns-default-zbhcm\" (UID: \"f6fc593a-b4b9-46ee-b12f-9ac4732dc69e\") " pod="openshift-dns/dns-default-zbhcm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.662948 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-mountpoint-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.664986 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cc69668d-b5d6-452a-99bf-b9a14c90a65d-stats-auth\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.665701 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693-node-bootstrap-token\") pod \"machine-config-server-gwfw8\" (UID: \"41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693\") " pod="openshift-machine-config-operator/machine-config-server-gwfw8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.666761 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc69668d-b5d6-452a-99bf-b9a14c90a65d-service-ca-bundle\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.685036 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/777fcbda-0e10-4a62-aa45-fa8670bae076-config\") pod \"service-ca-operator-777779d784-nflhv\" (UID: \"777fcbda-0e10-4a62-aa45-fa8670bae076\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.685509 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6fc593a-b4b9-46ee-b12f-9ac4732dc69e-config-volume\") pod \"dns-default-zbhcm\" (UID: \"f6fc593a-b4b9-46ee-b12f-9ac4732dc69e\") " pod="openshift-dns/dns-default-zbhcm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.692177 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7t7px\" (UID: \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\") " pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.697031 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-plugins-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.702167 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7308dfe6-f141-4f20-a966-075d17882333-config-volume\") pod \"collect-profiles-29403975-mz4zk\" (UID: \"7308dfe6-f141-4f20-a966-075d17882333\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.702276 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-csi-data-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.702386 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a053f34f-9565-4d7c-88bc-2edd1c08cb69-cert\") pod \"ingress-canary-57tr4\" (UID: \"a053f34f-9565-4d7c-88bc-2edd1c08cb69\") " pod="openshift-ingress-canary/ingress-canary-57tr4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.702997 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cc69668d-b5d6-452a-99bf-b9a14c90a65d-default-certificate\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.704368 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3fb2578f-5e34-45b4-bcae-4798bfba073a-profile-collector-cert\") pod \"catalog-operator-68c6474976-xvtnq\" (UID: \"3fb2578f-5e34-45b4-bcae-4798bfba073a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.704908 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7308dfe6-f141-4f20-a966-075d17882333-secret-volume\") pod \"collect-profiles-29403975-mz4zk\" (UID: \"7308dfe6-f141-4f20-a966-075d17882333\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.705522 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693-certs\") pod \"machine-config-server-gwfw8\" (UID: \"41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693\") " pod="openshift-machine-config-operator/machine-config-server-gwfw8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.706444 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-socket-dir\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.706868 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea7b78e0-1c47-45a1-b93c-bd7eba502143-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-sg6x8\" (UID: \"ea7b78e0-1c47-45a1-b93c-bd7eba502143\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.715522 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06-images\") pod \"machine-config-operator-74547568cd-prj6n\" (UID: \"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.726456 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/42e226ef-f2d6-4a0a-8d35-f9a095322a81-srv-cert\") pod \"olm-operator-6b444d44fb-pxg4x\" (UID: \"42e226ef-f2d6-4a0a-8d35-f9a095322a81\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.729893 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc"] Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.732555 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7t7px\" (UID: \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\") " pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.736453 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6tbp\" (UniqueName: \"kubernetes.io/projected/42e226ef-f2d6-4a0a-8d35-f9a095322a81-kube-api-access-p6tbp\") pod \"olm-operator-6b444d44fb-pxg4x\" (UID: \"42e226ef-f2d6-4a0a-8d35-f9a095322a81\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.742735 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zd4hv"] Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.745372 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: E1127 10:23:57.745901 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:23:58.245873042 +0000 UTC m=+142.832463806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.748018 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/42e226ef-f2d6-4a0a-8d35-f9a095322a81-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pxg4x\" (UID: \"42e226ef-f2d6-4a0a-8d35-f9a095322a81\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.749791 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvg48\" (UniqueName: \"kubernetes.io/projected/67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06-kube-api-access-gvg48\") pod \"machine-config-operator-74547568cd-prj6n\" (UID: \"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.751615 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06-proxy-tls\") pod \"machine-config-operator-74547568cd-prj6n\" (UID: \"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.751941 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5qtlg\" (UID: \"adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.752151 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3fb2578f-5e34-45b4-bcae-4798bfba073a-srv-cert\") pod \"catalog-operator-68c6474976-xvtnq\" (UID: \"3fb2578f-5e34-45b4-bcae-4798bfba073a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.754144 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nftbv\" (UniqueName: \"kubernetes.io/projected/a053f34f-9565-4d7c-88bc-2edd1c08cb69-kube-api-access-nftbv\") pod \"ingress-canary-57tr4\" (UID: \"a053f34f-9565-4d7c-88bc-2edd1c08cb69\") " pod="openshift-ingress-canary/ingress-canary-57tr4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.765873 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsddn\" (UniqueName: \"kubernetes.io/projected/777fcbda-0e10-4a62-aa45-fa8670bae076-kube-api-access-bsddn\") pod \"service-ca-operator-777779d784-nflhv\" (UID: \"777fcbda-0e10-4a62-aa45-fa8670bae076\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.793676 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2b8p\" (UniqueName: \"kubernetes.io/projected/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-kube-api-access-d2b8p\") pod \"marketplace-operator-79b997595-7t7px\" (UID: \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\") " pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.802158 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnbph\" (UniqueName: \"kubernetes.io/projected/41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693-kube-api-access-bnbph\") pod \"machine-config-server-gwfw8\" (UID: \"41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693\") " pod="openshift-machine-config-operator/machine-config-server-gwfw8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.826018 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-lzwl4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.841283 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgcdp\" (UniqueName: \"kubernetes.io/projected/9b0fd0cb-13fa-4f87-bfc3-defe908884d4-kube-api-access-wgcdp\") pod \"service-ca-9c57cc56f-28rbb\" (UID: \"9b0fd0cb-13fa-4f87-bfc3-defe908884d4\") " pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.846576 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:57 crc kubenswrapper[4806]: E1127 10:23:57.847051 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:23:58.347035272 +0000 UTC m=+142.933626026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.855986 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc"] Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.857594 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kpvd\" (UniqueName: \"kubernetes.io/projected/f6fc593a-b4b9-46ee-b12f-9ac4732dc69e-kube-api-access-9kpvd\") pod \"dns-default-zbhcm\" (UID: \"f6fc593a-b4b9-46ee-b12f-9ac4732dc69e\") " pod="openshift-dns/dns-default-zbhcm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.864431 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnn7k\" (UniqueName: \"kubernetes.io/projected/ea7b78e0-1c47-45a1-b93c-bd7eba502143-kube-api-access-nnn7k\") pod \"package-server-manager-789f6589d5-sg6x8\" (UID: \"ea7b78e0-1c47-45a1-b93c-bd7eba502143\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.871429 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.874644 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnjm4\" (UniqueName: \"kubernetes.io/projected/3c890acc-37e4-4838-92c5-ff9a72738519-kube-api-access-cnjm4\") pod \"migrator-59844c95c7-62tks\" (UID: \"3c890acc-37e4-4838-92c5-ff9a72738519\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-62tks" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.879508 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-42h4s"] Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.880332 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.890614 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-62tks" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.904567 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.905068 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.914462 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c"] Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.915993 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb2fx\" (UniqueName: \"kubernetes.io/projected/3fb2578f-5e34-45b4-bcae-4798bfba073a-kube-api-access-zb2fx\") pod \"catalog-operator-68c6474976-xvtnq\" (UID: \"3fb2578f-5e34-45b4-bcae-4798bfba073a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.917218 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.923842 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.938626 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-57tr4" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.951342 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9lgw\" (UniqueName: \"kubernetes.io/projected/cc69668d-b5d6-452a-99bf-b9a14c90a65d-kube-api-access-g9lgw\") pod \"router-default-5444994796-kpjrm\" (UID: \"cc69668d-b5d6-452a-99bf-b9a14c90a65d\") " pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.951364 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zbhcm" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.957978 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvj2s\" (UniqueName: \"kubernetes.io/projected/d3b66f29-69bb-4b08-b3bd-ce924dbe7d43-kube-api-access-hvj2s\") pod \"csi-hostpathplugin-z979k\" (UID: \"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43\") " pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.964063 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn"] Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.965998 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gwfw8" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.966505 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm"] Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.966670 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:57 crc kubenswrapper[4806]: E1127 10:23:57.967036 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:23:58.467023437 +0000 UTC m=+143.053614201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.971209 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7zwh\" (UniqueName: \"kubernetes.io/projected/16a0adb3-9386-417e-8a34-bdf429462e2b-kube-api-access-w7zwh\") pod \"packageserver-d55dfcdfc-wsjh5\" (UID: \"16a0adb3-9386-417e-8a34-bdf429462e2b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.997979 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-z979k" Nov 27 10:23:57 crc kubenswrapper[4806]: I1127 10:23:57.999795 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.001666 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-z4ndr"] Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.004769 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk"] Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.004818 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9"] Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.034247 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbs8q\" (UniqueName: \"kubernetes.io/projected/7308dfe6-f141-4f20-a966-075d17882333-kube-api-access-rbs8q\") pod \"collect-profiles-29403975-mz4zk\" (UID: \"7308dfe6-f141-4f20-a966-075d17882333\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.034290 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" event={"ID":"52480926-8957-4c79-a384-3e2d27f349af","Type":"ContainerStarted","Data":"56498e84f8a5392827af97523e48fa154107237605d184b3c0be53ac7dcbe5dd"} Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.034331 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" event={"ID":"52480926-8957-4c79-a384-3e2d27f349af","Type":"ContainerStarted","Data":"95386bc40ac8d4ec85cbcb8a3781d65a29ac052f7ae1c403b04dd9149cd06d44"} Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.036265 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4fdd\" (UniqueName: \"kubernetes.io/projected/adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce-kube-api-access-l4fdd\") pod \"control-plane-machine-set-operator-78cbb6b69f-5qtlg\" (UID: \"adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg" Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.038789 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" event={"ID":"a52b3a70-665d-49e5-a842-14ade4dcff1f","Type":"ContainerStarted","Data":"4505df16199f9530483bf2b30fe6aab719388b368264289cf510baba2c8d249a"} Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.075063 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:58 crc kubenswrapper[4806]: E1127 10:23:58.075392 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:23:58.575376849 +0000 UTC m=+143.161967603 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.085811 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l956q"] Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.095574 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9ppfs" event={"ID":"52dd6d6d-3a15-4467-9d82-37fd2fa5794b","Type":"ContainerStarted","Data":"8316b3f2b0d3ece449477846dfc8dd7fe4e355e795d49f5f8cc390d0ba62cd56"} Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.113064 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" event={"ID":"d26adb0c-6bf0-49f1-806b-06d96e333fc4","Type":"ContainerStarted","Data":"201f83dc3afc181bb3c95f87e3865cbc5b85c9bd72103b12a6fa8caaca5d80ad"} Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.136254 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6fqzk"] Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.153355 4806 generic.go:334] "Generic (PLEG): container finished" podID="65b0faa1-6054-43a4-ab15-e59791790337" containerID="18b335a9d5e3d3377f2a9db6efb311163766b04301b9e9099ca915d260b5d18b" exitCode=0 Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.153424 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" event={"ID":"65b0faa1-6054-43a4-ab15-e59791790337","Type":"ContainerDied","Data":"18b335a9d5e3d3377f2a9db6efb311163766b04301b9e9099ca915d260b5d18b"} Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.153454 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" event={"ID":"65b0faa1-6054-43a4-ab15-e59791790337","Type":"ContainerStarted","Data":"00d4992b316851fc1a75c2977fade1bac50d27899104c6a6dfc8e51c3f50f25d"} Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.159405 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.160777 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" event={"ID":"6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe","Type":"ContainerStarted","Data":"80ace9e2d2690adb4c6395a4ebae7df6ac56e7be626bda80c62b43918d4097aa"} Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.162716 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg" Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.162940 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" event={"ID":"325e82c6-eacf-4180-bc5f-9334536ebe16","Type":"ContainerStarted","Data":"46024a99ea6c55f28a25e17ee22985c00cfa9232e09b324ffc7cfc8d356f8da8"} Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.179040 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:58 crc kubenswrapper[4806]: E1127 10:23:58.180669 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:23:58.680648944 +0000 UTC m=+143.267239708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.188395 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2szrd"] Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.212149 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.236923 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.281022 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:58 crc kubenswrapper[4806]: E1127 10:23:58.281501 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:23:58.781438033 +0000 UTC m=+143.368028787 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.281758 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:58 crc kubenswrapper[4806]: E1127 10:23:58.283011 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:23:58.783002457 +0000 UTC m=+143.369593221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:58 crc kubenswrapper[4806]: W1127 10:23:58.325503 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfd928ce_640d_4b02_9d41_a81a12b3fbfc.slice/crio-89a15517e56ec66939df665561ceecac36e45c434e969c23ffa2dd070ba4855c WatchSource:0}: Error finding container 89a15517e56ec66939df665561ceecac36e45c434e969c23ffa2dd070ba4855c: Status 404 returned error can't find the container with id 89a15517e56ec66939df665561ceecac36e45c434e969c23ffa2dd070ba4855c Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.330145 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8jvb8"] Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.383119 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:58 crc kubenswrapper[4806]: E1127 10:23:58.383442 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:23:58.883427396 +0000 UTC m=+143.470018160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.484486 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:58 crc kubenswrapper[4806]: E1127 10:23:58.484960 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:23:58.984933555 +0000 UTC m=+143.571524499 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.559809 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r"] Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.594925 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:58 crc kubenswrapper[4806]: E1127 10:23:58.595896 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:23:59.095875689 +0000 UTC m=+143.682466453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.609333 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8"] Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.697120 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:58 crc kubenswrapper[4806]: E1127 10:23:58.697502 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:23:59.197489411 +0000 UTC m=+143.784080175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.709502 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8gdsx" podStartSLOduration=122.709484018 podStartE2EDuration="2m2.709484018s" podCreationTimestamp="2025-11-27 10:21:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:58.668618044 +0000 UTC m=+143.255208808" watchObservedRunningTime="2025-11-27 10:23:58.709484018 +0000 UTC m=+143.296074782" Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.798766 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:58 crc kubenswrapper[4806]: E1127 10:23:58.799358 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:23:59.299331991 +0000 UTC m=+143.885922755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:58 crc kubenswrapper[4806]: W1127 10:23:58.823242 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13187e9e_9b78_41ce_a2ff_7837dfd8eca6.slice/crio-323b5770a609da39ec62bbb9a97cf32c108d00c083350bf4a8d703b56e04b5bf WatchSource:0}: Error finding container 323b5770a609da39ec62bbb9a97cf32c108d00c083350bf4a8d703b56e04b5bf: Status 404 returned error can't find the container with id 323b5770a609da39ec62bbb9a97cf32c108d00c083350bf4a8d703b56e04b5bf Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.888094 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-r7wlb"] Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.900407 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:58 crc kubenswrapper[4806]: E1127 10:23:58.900876 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:23:59.400860531 +0000 UTC m=+143.987451295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:58 crc kubenswrapper[4806]: I1127 10:23:58.908539 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts"] Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.001313 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:59 crc kubenswrapper[4806]: E1127 10:23:59.001783 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:23:59.501748803 +0000 UTC m=+144.088339567 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.063187 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lzwl4"] Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.104628 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:59 crc kubenswrapper[4806]: E1127 10:23:59.104983 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:23:59.60495738 +0000 UTC m=+144.191548144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.112807 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x"] Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.208857 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:59 crc kubenswrapper[4806]: E1127 10:23:59.209286 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:23:59.709220707 +0000 UTC m=+144.295811471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.241408 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8"] Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.251800 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-42h4s" event={"ID":"d52f4ad9-009f-41cf-a779-2a6515a2fbb4","Type":"ContainerStarted","Data":"64ea239bcb365392bc8f0002da76c2efb6b03f6ba2f73b302bac883e77364ad0"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.252129 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-42h4s" event={"ID":"d52f4ad9-009f-41cf-a779-2a6515a2fbb4","Type":"ContainerStarted","Data":"7137db5ea3a076c79c3371a29522dbe2f556d387fd36c87318c1d8f87e22cd21"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.252149 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.266944 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" event={"ID":"f893e28e-fb92-45b2-8c43-05ff99bf7d32","Type":"ContainerStarted","Data":"bc743a67a7e3ce9e010156a5b344edb316cd3220a1b8b9553011c51a85e20661"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.271090 4806 patch_prober.go:28] interesting pod/console-operator-58897d9998-42h4s container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.271128 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-42h4s" podUID="d52f4ad9-009f-41cf-a779-2a6515a2fbb4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.276688 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l956q" event={"ID":"0639fe21-6b99-4fad-be7b-6bd998c3472c","Type":"ContainerStarted","Data":"017278a159ddfb26c4b4693cccca6595d5987738690262a9d9f06a9f3620f1e6"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.278926 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7t7px"] Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.300923 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" event={"ID":"c225ab90-51ee-400f-8902-774ad6270322","Type":"ContainerStarted","Data":"0d681d5b8d6670489ba4010f72e34e5ec25e1f554e304f8802673b13a0b45fc3"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.315709 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:59 crc kubenswrapper[4806]: E1127 10:23:59.316067 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:23:59.816055475 +0000 UTC m=+144.402646239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.323087 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" event={"ID":"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f","Type":"ContainerStarted","Data":"a67bcd61d986e56f9ded14e5b71955e79bef00bd9bf00ad29f44930ab6420ac4"} Nov 27 10:23:59 crc kubenswrapper[4806]: W1127 10:23:59.339845 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9ef04e5_e987_4457_bd8b_d1bc01cb67ac.slice/crio-fa9e55c611221599ab3f925b702161b46c7095464297bd9797fe14ed6fef832e WatchSource:0}: Error finding container fa9e55c611221599ab3f925b702161b46c7095464297bd9797fe14ed6fef832e: Status 404 returned error can't find the container with id fa9e55c611221599ab3f925b702161b46c7095464297bd9797fe14ed6fef832e Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.345369 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh"] Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.380224 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-kpjrm" event={"ID":"cc69668d-b5d6-452a-99bf-b9a14c90a65d","Type":"ContainerStarted","Data":"1ae27793012771a2a3aeb200df3994df7c46a13df0168c5a3b5a1e602a304526"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.397989 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-9ppfs" Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.398023 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9ppfs" event={"ID":"52dd6d6d-3a15-4467-9d82-37fd2fa5794b","Type":"ContainerStarted","Data":"52055cedb9d6dd2ab8b23ffdfc7a72b2ae59ef6385d43a6da32aaf9ba55df70b"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.418165 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:59 crc kubenswrapper[4806]: E1127 10:23:59.419937 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:23:59.919791337 +0000 UTC m=+144.506382101 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.435484 4806 patch_prober.go:28] interesting pod/downloads-7954f5f757-9ppfs container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.435541 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9ppfs" podUID="52dd6d6d-3a15-4467-9d82-37fd2fa5794b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.446260 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" event={"ID":"e60de411-4929-4a3a-a5ed-1071da7ed16a","Type":"ContainerStarted","Data":"04996a302f9cddd1b69f44131503d73e946b4dfac08e5e4f3b8c62c8c8638e92"} Nov 27 10:23:59 crc kubenswrapper[4806]: W1127 10:23:59.471322 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea7b78e0_1c47_45a1_b93c_bd7eba502143.slice/crio-280d20301171399d00529bde8ce0cbda0932f49d60a1ee08f09be78a5786e207 WatchSource:0}: Error finding container 280d20301171399d00529bde8ce0cbda0932f49d60a1ee08f09be78a5786e207: Status 404 returned error can't find the container with id 280d20301171399d00529bde8ce0cbda0932f49d60a1ee08f09be78a5786e207 Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.476416 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6fqzk" event={"ID":"97042267-61ca-4864-a50a-60587abc7619","Type":"ContainerStarted","Data":"46651de76bba572654753c832084eb28b0db7ecba5777cad0076c1426de15951"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.486072 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9" event={"ID":"a3ce016c-a670-46a3-9254-03fea2c6568b","Type":"ContainerStarted","Data":"9c6a0181c7b643191171180183b9a2535f03d699413e56eb2802dbeb39878834"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.507828 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-9ppfs" podStartSLOduration=122.507810959 podStartE2EDuration="2m2.507810959s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:59.465154156 +0000 UTC m=+144.051744920" watchObservedRunningTime="2025-11-27 10:23:59.507810959 +0000 UTC m=+144.094401723" Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.508197 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-42h4s" podStartSLOduration=122.50819243 podStartE2EDuration="2m2.50819243s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:23:59.507647414 +0000 UTC m=+144.094238178" watchObservedRunningTime="2025-11-27 10:23:59.50819243 +0000 UTC m=+144.094783194" Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.508538 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" event={"ID":"6ccc4b13-3a1d-42f8-8b3b-9fdb672be8fe","Type":"ContainerStarted","Data":"a4f4b8a2f9a4336aa2336800fdd5649a662580a315302a4438c88f0bfcb3dd7c"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.519821 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:59 crc kubenswrapper[4806]: E1127 10:23:59.520178 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:00.020166425 +0000 UTC m=+144.606757189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.525178 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-62tks"] Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.544797 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" event={"ID":"d26adb0c-6bf0-49f1-806b-06d96e333fc4","Type":"ContainerStarted","Data":"3cb0ecf188166feb44845bd0f061fdb7bcc635aa5565127c157be9ca7908d97e"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.550629 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-28rbb"] Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.568122 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" event={"ID":"325e82c6-eacf-4180-bc5f-9334536ebe16","Type":"ContainerStarted","Data":"83069f30439eda3511df0a1764fa47dcbe31174dd08c23e273f446d8065a745e"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.574915 4806 generic.go:334] "Generic (PLEG): container finished" podID="a52b3a70-665d-49e5-a842-14ade4dcff1f" containerID="155ad7c622fe69c4c9a8928df3834a9fc62a390daff31d9af1d0d5ffb6354bf7" exitCode=0 Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.574963 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" event={"ID":"a52b3a70-665d-49e5-a842-14ade4dcff1f","Type":"ContainerDied","Data":"155ad7c622fe69c4c9a8928df3834a9fc62a390daff31d9af1d0d5ffb6354bf7"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.596254 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" event={"ID":"dfd928ce-640d-4b02-9d41-a81a12b3fbfc","Type":"ContainerStarted","Data":"89a15517e56ec66939df665561ceecac36e45c434e969c23ffa2dd070ba4855c"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.602169 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" event={"ID":"f25c1d0d-db4f-415b-a7ea-4d1b98e84819","Type":"ContainerStarted","Data":"e884a00758cec961e60a78ef1a736cbeb5353c04ba76cadbff6d1f22c571d0cd"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.613685 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" event={"ID":"13187e9e-9b78-41ce-a2ff-7837dfd8eca6","Type":"ContainerStarted","Data":"323b5770a609da39ec62bbb9a97cf32c108d00c083350bf4a8d703b56e04b5bf"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.616744 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" event={"ID":"93b844fb-455e-4a0a-9d75-52f1278df25c","Type":"ContainerStarted","Data":"5809551bcc7a9b9acfef46ea9679c7d4d7dcac9fb423ae77e0204b3c0919bdd8"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.621015 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:59 crc kubenswrapper[4806]: E1127 10:23:59.621334 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:00.121292874 +0000 UTC m=+144.707883638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:59 crc kubenswrapper[4806]: W1127 10:23:59.622019 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b0fd0cb_13fa_4f87_bfc3_defe908884d4.slice/crio-31f9b5285990c33d34fa968d34fded14d310d652ee0772a27db4a829099c43a2 WatchSource:0}: Error finding container 31f9b5285990c33d34fa968d34fded14d310d652ee0772a27db4a829099c43a2: Status 404 returned error can't find the container with id 31f9b5285990c33d34fa968d34fded14d310d652ee0772a27db4a829099c43a2 Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.622488 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:59 crc kubenswrapper[4806]: E1127 10:23:59.623205 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:00.123189657 +0000 UTC m=+144.709780421 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.624728 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gwfw8" event={"ID":"41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693","Type":"ContainerStarted","Data":"bf925775d3904fdfddd1a10b3603175ff84c75e75ef416b912b1f621653a155b"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.630577 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-r7wlb" event={"ID":"e858c9d3-3f1c-47db-a0c8-f0efac1ed114","Type":"ContainerStarted","Data":"0d50eddf77e6bd4ca7664a933338719cf89fa4aae3d970d132fc1f491b3d80a2"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.668853 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" event={"ID":"65b0faa1-6054-43a4-ab15-e59791790337","Type":"ContainerStarted","Data":"958d1ff3a42c9e4ae68026a45b5c8a5ff92f9ca6a0757c75523b796039b85d4b"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.669731 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.674558 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" event={"ID":"e623ba7e-e936-4ab0-a825-6139b2b33fe4","Type":"ContainerStarted","Data":"668bfb75b0b5506a3f06f85b7b3332828ed0711b6a14769a8f4b07bc21e9ae44"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.699144 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" event={"ID":"47d1375e-f8bc-4406-8692-fe9aecb0e394","Type":"ContainerStarted","Data":"71d5dc359b14df4f742809a30c2377adc4c827b9bd9965dfa03bf2a5c7feea25"} Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.724976 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:59 crc kubenswrapper[4806]: E1127 10:23:59.727104 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:00.227082933 +0000 UTC m=+144.813673697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.786820 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-57tr4"] Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.827485 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:23:59 crc kubenswrapper[4806]: E1127 10:23:59.828728 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:00.328716586 +0000 UTC m=+144.915307360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.839370 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zbhcm"] Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.844686 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nflhv"] Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.853313 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z979k"] Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.910057 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq"] Nov 27 10:23:59 crc kubenswrapper[4806]: I1127 10:23:59.929306 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:23:59 crc kubenswrapper[4806]: E1127 10:23:59.930098 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:00.430067041 +0000 UTC m=+145.016657805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.001069 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk"] Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.015882 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5"] Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.030394 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg"] Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.034374 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" podStartSLOduration=124.034353178 podStartE2EDuration="2m4.034353178s" podCreationTimestamp="2025-11-27 10:21:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:00.018851955 +0000 UTC m=+144.605442719" watchObservedRunningTime="2025-11-27 10:24:00.034353178 +0000 UTC m=+144.620943942" Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.063002 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:00 crc kubenswrapper[4806]: E1127 10:24:00.064007 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:00.563988878 +0000 UTC m=+145.150579632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.069743 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-dqjqc" podStartSLOduration=123.069721408 podStartE2EDuration="2m3.069721408s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:00.069048639 +0000 UTC m=+144.655639403" watchObservedRunningTime="2025-11-27 10:24:00.069721408 +0000 UTC m=+144.656312172" Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.152634 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n"] Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.157915 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" podStartSLOduration=123.157887115 podStartE2EDuration="2m3.157887115s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:00.156899837 +0000 UTC m=+144.743490601" watchObservedRunningTime="2025-11-27 10:24:00.157887115 +0000 UTC m=+144.744477879" Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.172170 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:00 crc kubenswrapper[4806]: E1127 10:24:00.172813 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:00.672776301 +0000 UTC m=+145.259367065 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.273983 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:00 crc kubenswrapper[4806]: E1127 10:24:00.274432 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:00.774421514 +0000 UTC m=+145.361012278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.375003 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:00 crc kubenswrapper[4806]: E1127 10:24:00.375298 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:00.875273585 +0000 UTC m=+145.461864349 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.375365 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:00 crc kubenswrapper[4806]: E1127 10:24:00.375888 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:00.875869982 +0000 UTC m=+145.462460746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.477844 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:00 crc kubenswrapper[4806]: E1127 10:24:00.479382 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:00.979359147 +0000 UTC m=+145.565949911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.580464 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:00 crc kubenswrapper[4806]: E1127 10:24:00.581052 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.081023261 +0000 UTC m=+145.667614025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.681747 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:00 crc kubenswrapper[4806]: E1127 10:24:00.681947 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.181916294 +0000 UTC m=+145.768507058 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.682451 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:00 crc kubenswrapper[4806]: E1127 10:24:00.682818 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.182810578 +0000 UTC m=+145.769401342 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.785999 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:00 crc kubenswrapper[4806]: E1127 10:24:00.786450 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.286404236 +0000 UTC m=+145.872995000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: E1127 10:24:00.788218 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.288200807 +0000 UTC m=+145.874791571 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.804846 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.852836 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" event={"ID":"ea7b78e0-1c47-45a1-b93c-bd7eba502143","Type":"ContainerStarted","Data":"280d20301171399d00529bde8ce0cbda0932f49d60a1ee08f09be78a5786e207"} Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.890680 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-z4ndr" event={"ID":"93b844fb-455e-4a0a-9d75-52f1278df25c","Type":"ContainerStarted","Data":"874ed570e6b80ea286ae4601975b5e53c25c2271020d3d4cc4fc60997c160092"} Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.896464 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gwfw8" event={"ID":"41ba1b82-4ca5-4ef2-ac24-7f5ae35e6693","Type":"ContainerStarted","Data":"3b65647cbbe34f349bf7bf2586dec55145cb23261900217a19de9895c2ba5903"} Nov 27 10:24:00 crc kubenswrapper[4806]: E1127 10:24:00.909830 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.409817459 +0000 UTC m=+145.996408223 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.909764 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.910048 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:00 crc kubenswrapper[4806]: E1127 10:24:00.910871 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.410860947 +0000 UTC m=+145.997451711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.913026 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l956q" event={"ID":"0639fe21-6b99-4fad-be7b-6bd998c3472c","Type":"ContainerStarted","Data":"6141b54ffe61baaec35a3bab6aed1a55f0d15f2e5259d6178f1658b6e7522320"} Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.913874 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.919600 4806 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-l956q container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" start-of-body= Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.919665 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-l956q" podUID="0639fe21-6b99-4fad-be7b-6bd998c3472c" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.924200 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" event={"ID":"e623ba7e-e936-4ab0-a825-6139b2b33fe4","Type":"ContainerStarted","Data":"14a152f55515271f96d16b8f9aefc07685a15bc1925a40c53527467031ba2b4a"} Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.945400 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" event={"ID":"777fcbda-0e10-4a62-aa45-fa8670bae076","Type":"ContainerStarted","Data":"8156534ee02a9836459bca10bd2be91c8c6cff375e12830b39fb466e9a617479"} Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.955383 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" event={"ID":"42e226ef-f2d6-4a0a-8d35-f9a095322a81","Type":"ContainerStarted","Data":"8bf14fd2466725ddfbbd7b58c584f689a410df83773ad1764689c40f89711f27"} Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.955416 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" event={"ID":"42e226ef-f2d6-4a0a-8d35-f9a095322a81","Type":"ContainerStarted","Data":"2fd14d5c08c235d0d9ca0b2dba55386ecca90438ec6d07a031f7daacc6f87d24"} Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.956628 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.961131 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-gwfw8" podStartSLOduration=6.961113683 podStartE2EDuration="6.961113683s" podCreationTimestamp="2025-11-27 10:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:00.932830832 +0000 UTC m=+145.519421596" watchObservedRunningTime="2025-11-27 10:24:00.961113683 +0000 UTC m=+145.547704447" Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.963306 4806 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-pxg4x container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.963341 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" podUID="42e226ef-f2d6-4a0a-8d35-f9a095322a81" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Nov 27 10:24:00 crc kubenswrapper[4806]: I1127 10:24:00.968654 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" event={"ID":"13187e9e-9b78-41ce-a2ff-7837dfd8eca6","Type":"ContainerStarted","Data":"4cd034eb576a0fe344f973f6258bc6994817768727fefd8360cbd5f126e7160c"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.012378 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.012768 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.512750128 +0000 UTC m=+146.099340892 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.013106 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.015520 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kb5rn" podStartSLOduration=124.015482035 podStartE2EDuration="2m4.015482035s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:00.963627364 +0000 UTC m=+145.550218118" watchObservedRunningTime="2025-11-27 10:24:01.015482035 +0000 UTC m=+145.602072799" Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.029822 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.529802895 +0000 UTC m=+146.116393659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.038799 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-l956q" podStartSLOduration=125.038756186 podStartE2EDuration="2m5.038756186s" podCreationTimestamp="2025-11-27 10:21:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.001546295 +0000 UTC m=+145.588137079" watchObservedRunningTime="2025-11-27 10:24:01.038756186 +0000 UTC m=+145.625346950" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.069048 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fx54r" podStartSLOduration=124.069031093 podStartE2EDuration="2m4.069031093s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.066873602 +0000 UTC m=+145.653464366" watchObservedRunningTime="2025-11-27 10:24:01.069031093 +0000 UTC m=+145.655621857" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.115843 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" event={"ID":"4d69e866-000e-4748-9953-4d5f007f2669","Type":"ContainerStarted","Data":"ffdcbe37932d66789b840ead0c441865b41d954b271fc78f7731ba613566d8ab"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.119931 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.120415 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.620390319 +0000 UTC m=+146.206981083 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.120568 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.120707 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" podStartSLOduration=124.120695827 podStartE2EDuration="2m4.120695827s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.099405102 +0000 UTC m=+145.685995866" watchObservedRunningTime="2025-11-27 10:24:01.120695827 +0000 UTC m=+145.707286591" Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.121984 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.621965964 +0000 UTC m=+146.208556728 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.195044 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9" event={"ID":"a3ce016c-a670-46a3-9254-03fea2c6568b","Type":"ContainerStarted","Data":"61cad8518b24b69b0f577239b057ebe2269607ebdc8770f271bd3ef8bfece9a5"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.214633 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zbhcm" event={"ID":"f6fc593a-b4b9-46ee-b12f-9ac4732dc69e","Type":"ContainerStarted","Data":"b19d1ba6a9b0f0a342ffd5229eddfed842484613b58672153e6443766d7106a7"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.223411 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.226829 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.726800976 +0000 UTC m=+146.313391740 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.227816 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.228196 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.728183055 +0000 UTC m=+146.314773819 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.256951 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" event={"ID":"d26adb0c-6bf0-49f1-806b-06d96e333fc4","Type":"ContainerStarted","Data":"f3ba3ecb0c0633cab8eb02152c8175f490cf33dc3db298d35b7eaeba70765ed4"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.269143 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" event={"ID":"dfd928ce-640d-4b02-9d41-a81a12b3fbfc","Type":"ContainerStarted","Data":"88784f783a81f6f615b47abf4bab6fc4290e752297efcf5b66c9ea897f0c621b"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.271899 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.274564 4806 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-2szrd container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.274618 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" podUID="dfd928ce-640d-4b02-9d41-a81a12b3fbfc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.286600 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-sqqv6" podStartSLOduration=125.286572647 podStartE2EDuration="2m5.286572647s" podCreationTimestamp="2025-11-27 10:21:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.285395965 +0000 UTC m=+145.871986729" watchObservedRunningTime="2025-11-27 10:24:01.286572647 +0000 UTC m=+145.873163411" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.313650 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-62tks" event={"ID":"3c890acc-37e4-4838-92c5-ff9a72738519","Type":"ContainerStarted","Data":"ef17ed7579889026fe29839bf2f4c0a89f93110cf711eec328e251d77c2534b5"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.323275 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-57tr4" event={"ID":"a053f34f-9565-4d7c-88bc-2edd1c08cb69","Type":"ContainerStarted","Data":"0936f4d34197869255f103216ebcf4297e9843c5e289b04445d823d0130376ee"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.324795 4806 generic.go:334] "Generic (PLEG): container finished" podID="47d1375e-f8bc-4406-8692-fe9aecb0e394" containerID="264d894f663007c58f65158e814d69be0ffd54586ea8b8a24da77ba9dab33116" exitCode=0 Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.324844 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" event={"ID":"47d1375e-f8bc-4406-8692-fe9aecb0e394","Type":"ContainerDied","Data":"264d894f663007c58f65158e814d69be0ffd54586ea8b8a24da77ba9dab33116"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.328567 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.328928 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.828910163 +0000 UTC m=+146.415500917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.331216 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" event={"ID":"7ce6a8cd-9071-4c23-a36f-9af525e7f04e","Type":"ContainerStarted","Data":"ae3e40924632603c92e2c352997042f2f40c55caaf58bc813a613c6130eb2a7e"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.332195 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.361476 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" podStartSLOduration=124.361463023 podStartE2EDuration="2m4.361463023s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.36065465 +0000 UTC m=+145.947245414" watchObservedRunningTime="2025-11-27 10:24:01.361463023 +0000 UTC m=+145.948053777" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.362422 4806 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7t7px container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.362482 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" podUID="7ce6a8cd-9071-4c23-a36f-9af525e7f04e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.364789 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" event={"ID":"c225ab90-51ee-400f-8902-774ad6270322","Type":"ContainerStarted","Data":"d3a6f92bfd227606022ec3dda6b7cb5b393fef3989021254372bf0e9a3974b5b"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.376258 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lzwl4" event={"ID":"f9ef04e5-e987-4457-bd8b-d1bc01cb67ac","Type":"ContainerStarted","Data":"fa9e55c611221599ab3f925b702161b46c7095464297bd9797fe14ed6fef832e"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.377407 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" event={"ID":"16a0adb3-9386-417e-8a34-bdf429462e2b","Type":"ContainerStarted","Data":"0b9a253cb0d9fb3c889f21c61f3b0302be81484ac817b3615340e208b9650061"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.378389 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z979k" event={"ID":"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43","Type":"ContainerStarted","Data":"8f3ea4206d78b0944748b3d662903c679fd14582d266ee057c92a6a9dccb0358"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.385176 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" event={"ID":"06f03ec3-796d-442f-8adc-93aa22466426","Type":"ContainerStarted","Data":"dcc9d074a24c80d64cb24d04e611b90c925f302be5630a356d874c5a80bdbab4"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.388676 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-kpjrm" event={"ID":"cc69668d-b5d6-452a-99bf-b9a14c90a65d","Type":"ContainerStarted","Data":"7191d227ec20eb7e9e64a0b0435835414d52dcc13486540ac3affd0504b7e287"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.401282 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" event={"ID":"325e82c6-eacf-4180-bc5f-9334536ebe16","Type":"ContainerStarted","Data":"ba6895de88525022052822d1067c181e8e4201fb578dea3019e9db927c8072cb"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.409405 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" event={"ID":"7308dfe6-f141-4f20-a966-075d17882333","Type":"ContainerStarted","Data":"3d89ba4efb288e12349ee7d6e13a7001021ceca7eae9fb2e851e093a9a19da3e"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.411003 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" event={"ID":"f893e28e-fb92-45b2-8c43-05ff99bf7d32","Type":"ContainerStarted","Data":"0c2837a8fa2d08fa1169bc6f4baa963e6ebbd7f7ffee3c6fd952fb4f267f294c"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.426888 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" event={"ID":"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f","Type":"ContainerStarted","Data":"ea22f82fa7fcdf63523429c14e16ee4ed5a354fc2536a317d148606319f95d4c"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.427812 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.430912 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.433480 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:01.933468017 +0000 UTC m=+146.520058771 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.438281 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6fqzk" event={"ID":"97042267-61ca-4864-a50a-60587abc7619","Type":"ContainerStarted","Data":"10fb598b21604c336a6376d0e991967cd9fb1d34a7a4f7efe74b15d6921b0c8e"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.442721 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" event={"ID":"9b0fd0cb-13fa-4f87-bfc3-defe908884d4","Type":"ContainerStarted","Data":"31f9b5285990c33d34fa968d34fded14d310d652ee0772a27db4a829099c43a2"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.446964 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" podStartSLOduration=124.446952385 podStartE2EDuration="2m4.446952385s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.38174713 +0000 UTC m=+145.968337904" watchObservedRunningTime="2025-11-27 10:24:01.446952385 +0000 UTC m=+146.033543149" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.454183 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" event={"ID":"3fb2578f-5e34-45b4-bcae-4798bfba073a","Type":"ContainerStarted","Data":"df7b1cd4760f8a7a9a1536444ac8b1b646e4568db36c77dd1633f3c38fcaea9c"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.457760 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" event={"ID":"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06","Type":"ContainerStarted","Data":"14759fbf49881a6815e6c97e3dead1d09e47f7962a392e48b38c5be65d7bc4d5"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.467404 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg" event={"ID":"adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce","Type":"ContainerStarted","Data":"5c0b6c0e00bb70ed7f65a543d05ac1b085ce3e24dc3cf60d4324baaefb53ef13"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.481273 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" event={"ID":"f25c1d0d-db4f-415b-a7ea-4d1b98e84819","Type":"ContainerStarted","Data":"614d2eb6011ad1b02210c4b7fd69108dc6f72468bf5b096afd41b08f31e6de7d"} Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.489237 4806 patch_prober.go:28] interesting pod/downloads-7954f5f757-9ppfs container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.489308 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9ppfs" podUID="52dd6d6d-3a15-4467-9d82-37fd2fa5794b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.505175 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-pppcb" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.533186 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.534756 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:02.03473929 +0000 UTC m=+146.621330054 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.563933 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" podStartSLOduration=124.563913526 podStartE2EDuration="2m4.563913526s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.501065568 +0000 UTC m=+146.087656332" watchObservedRunningTime="2025-11-27 10:24:01.563913526 +0000 UTC m=+146.150504290" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.635132 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.636865 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:02.136852266 +0000 UTC m=+146.723443030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.645765 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" podStartSLOduration=124.645750006 podStartE2EDuration="2m4.645750006s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.56621314 +0000 UTC m=+146.152803904" watchObservedRunningTime="2025-11-27 10:24:01.645750006 +0000 UTC m=+146.232340770" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.739053 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.739335 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:02.239318713 +0000 UTC m=+146.825909477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.739448 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.739706 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:02.239698143 +0000 UTC m=+146.826288907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.761535 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-zd4hv" podStartSLOduration=124.761518454 podStartE2EDuration="2m4.761518454s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.680076986 +0000 UTC m=+146.266667750" watchObservedRunningTime="2025-11-27 10:24:01.761518454 +0000 UTC m=+146.348109218" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.762938 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-6fqzk" podStartSLOduration=124.76293265300001 podStartE2EDuration="2m4.762932653s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.760710442 +0000 UTC m=+146.347301206" watchObservedRunningTime="2025-11-27 10:24:01.762932653 +0000 UTC m=+146.349523417" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.840150 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.840618 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:02.340604016 +0000 UTC m=+146.927194770 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.867266 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-678qm" podStartSLOduration=124.867250492 podStartE2EDuration="2m4.867250492s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.801739109 +0000 UTC m=+146.388329873" watchObservedRunningTime="2025-11-27 10:24:01.867250492 +0000 UTC m=+146.453841256" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.901448 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-kpjrm" podStartSLOduration=124.901434188 podStartE2EDuration="2m4.901434188s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.869388461 +0000 UTC m=+146.455979225" watchObservedRunningTime="2025-11-27 10:24:01.901434188 +0000 UTC m=+146.488024952" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.902073 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" podStartSLOduration=124.902067445 podStartE2EDuration="2m4.902067445s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.89790923 +0000 UTC m=+146.484499994" watchObservedRunningTime="2025-11-27 10:24:01.902067445 +0000 UTC m=+146.488658209" Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.944879 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:01 crc kubenswrapper[4806]: E1127 10:24:01.945356 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:02.445340416 +0000 UTC m=+147.031931180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:01 crc kubenswrapper[4806]: I1127 10:24:01.965708 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwk9c" podStartSLOduration=124.965684645 podStartE2EDuration="2m4.965684645s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:01.964287466 +0000 UTC m=+146.550878230" watchObservedRunningTime="2025-11-27 10:24:01.965684645 +0000 UTC m=+146.552275409" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.046647 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:02 crc kubenswrapper[4806]: E1127 10:24:02.048283 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:02.547458763 +0000 UTC m=+147.134049527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.160537 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.160964 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:24:02 crc kubenswrapper[4806]: E1127 10:24:02.162198 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:02.662170902 +0000 UTC m=+147.248761666 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.197852 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:02 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:02 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:02 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.197932 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.255790 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.262185 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:02 crc kubenswrapper[4806]: E1127 10:24:02.262376 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:02.762351844 +0000 UTC m=+147.348942608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.262691 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:02 crc kubenswrapper[4806]: E1127 10:24:02.263159 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:02.763146286 +0000 UTC m=+147.349737050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.366921 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:02 crc kubenswrapper[4806]: E1127 10:24:02.367760 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:02.867731632 +0000 UTC m=+147.454322396 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.469493 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:02 crc kubenswrapper[4806]: E1127 10:24:02.470047 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:02.970032524 +0000 UTC m=+147.556623288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.483300 4806 patch_prober.go:28] interesting pod/console-operator-58897d9998-42h4s container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.483348 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-42h4s" podUID="d52f4ad9-009f-41cf-a779-2a6515a2fbb4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.539543 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lzwl4" event={"ID":"f9ef04e5-e987-4457-bd8b-d1bc01cb67ac","Type":"ContainerStarted","Data":"ad0da385a7a603164c5f29b41200218f9499501be74c3658d64d6bb84fa770ae"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.550970 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" event={"ID":"16a0adb3-9386-417e-8a34-bdf429462e2b","Type":"ContainerStarted","Data":"df4de010628d98483cb13596cab3a6a4587718998cd95f64bdafb4a188cee2a1"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.551678 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.557881 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zbhcm" event={"ID":"f6fc593a-b4b9-46ee-b12f-9ac4732dc69e","Type":"ContainerStarted","Data":"338ee2a4c8a942001d98c46951d51980da04d4042311aa2938f6c28fc18354e6"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.566976 4806 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wsjh5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" start-of-body= Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.567057 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" podUID="16a0adb3-9386-417e-8a34-bdf429462e2b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.570560 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:02 crc kubenswrapper[4806]: E1127 10:24:02.571731 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.071699828 +0000 UTC m=+147.658290612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.583010 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" event={"ID":"a52b3a70-665d-49e5-a842-14ade4dcff1f","Type":"ContainerStarted","Data":"b2170626fe7118e780483282a53bf48ee518beadc903c2957784886ae88d92d2"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.619428 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" event={"ID":"ea7b78e0-1c47-45a1-b93c-bd7eba502143","Type":"ContainerStarted","Data":"f402e1dc6d7185f051499e72cfd2b2809ff4856b3cdc15b3c8ad9918409d2d79"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.619508 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" event={"ID":"ea7b78e0-1c47-45a1-b93c-bd7eba502143","Type":"ContainerStarted","Data":"33a9a9e0577a13d9f200c981c2469fe2f50c52206c009b046e1626e4e9b31cf7"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.620893 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.665729 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" podStartSLOduration=125.665699567 podStartE2EDuration="2m5.665699567s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:02.592756607 +0000 UTC m=+147.179347371" watchObservedRunningTime="2025-11-27 10:24:02.665699567 +0000 UTC m=+147.252290351" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.666930 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" podStartSLOduration=125.666919451 podStartE2EDuration="2m5.666919451s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:02.665690897 +0000 UTC m=+147.252281661" watchObservedRunningTime="2025-11-27 10:24:02.666919451 +0000 UTC m=+147.253510215" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.674383 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:02 crc kubenswrapper[4806]: E1127 10:24:02.676252 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.176207101 +0000 UTC m=+147.762797865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.686597 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-62tks" event={"ID":"3c890acc-37e4-4838-92c5-ff9a72738519","Type":"ContainerStarted","Data":"67bf424edf301e72edeb086f01d522bc294fd039363e4489aa88ae23848ee06c"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.686668 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-62tks" event={"ID":"3c890acc-37e4-4838-92c5-ff9a72738519","Type":"ContainerStarted","Data":"cda40d870fdd567431d87b3a1f900c80c203e94f00062bb9b30849fedb846d05"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.721730 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-62tks" podStartSLOduration=125.721712864 podStartE2EDuration="2m5.721712864s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:02.716616692 +0000 UTC m=+147.303207456" watchObservedRunningTime="2025-11-27 10:24:02.721712864 +0000 UTC m=+147.308303628" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.730841 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-57tr4" event={"ID":"a053f34f-9565-4d7c-88bc-2edd1c08cb69","Type":"ContainerStarted","Data":"640deb6a7a5550d46087306a64b3a537ff924c501811963da51dd8ec6c362ce0"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.738595 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" event={"ID":"e60de411-4929-4a3a-a5ed-1071da7ed16a","Type":"ContainerStarted","Data":"fc615262e7b793049562e03a63e438cf4dd3fdf93792e73be047f38f74cec203"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.763469 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" event={"ID":"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06","Type":"ContainerStarted","Data":"66fccad8591aed92ac5103ffb3f97139015e288af0d79469dc516f5e2e90f49b"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.768015 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg" event={"ID":"adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce","Type":"ContainerStarted","Data":"bacfe60d72fecb3a06c2391e8c8ab31b97d4792db2fb70380f20b96fdb7e922b"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.775397 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:02 crc kubenswrapper[4806]: E1127 10:24:02.777177 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.277141955 +0000 UTC m=+147.863732769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.781926 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-57tr4" podStartSLOduration=8.781910178 podStartE2EDuration="8.781910178s" podCreationTimestamp="2025-11-27 10:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:02.781309361 +0000 UTC m=+147.367900125" watchObservedRunningTime="2025-11-27 10:24:02.781910178 +0000 UTC m=+147.368500942" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.782534 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-r7wlb" event={"ID":"e858c9d3-3f1c-47db-a0c8-f0efac1ed114","Type":"ContainerStarted","Data":"bae4f6ecd758e8f389f736f290331c7e176293b23bfd81e9b9ddaa4614fbd9d8"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.808848 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" event={"ID":"3fb2578f-5e34-45b4-bcae-4798bfba073a","Type":"ContainerStarted","Data":"4dc8c5288a8e45f20bfe181f05f966dcc99f7b78dc6504ea053c01394fd12c8b"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.809166 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.815790 4806 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-xvtnq container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.815849 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" podUID="3fb2578f-5e34-45b4-bcae-4798bfba073a" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.857555 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" event={"ID":"f893e28e-fb92-45b2-8c43-05ff99bf7d32","Type":"ContainerStarted","Data":"69f5ee06e30ff6372de58df1042ba36761c5fca8b79d69071473b8adac315235"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.861056 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-8jvb8" podStartSLOduration=125.861031821 podStartE2EDuration="2m5.861031821s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:02.858375258 +0000 UTC m=+147.444966022" watchObservedRunningTime="2025-11-27 10:24:02.861031821 +0000 UTC m=+147.447622585" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.868277 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" event={"ID":"47d1375e-f8bc-4406-8692-fe9aecb0e394","Type":"ContainerStarted","Data":"367629354b2292a325b7d9613026594f61a88939048df47c36be0358b2da160f"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.878767 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:02 crc kubenswrapper[4806]: E1127 10:24:02.880461 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.380443574 +0000 UTC m=+147.967034338 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.882596 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" event={"ID":"7308dfe6-f141-4f20-a966-075d17882333","Type":"ContainerStarted","Data":"535ec0a3c11c7d698bab89bdd18508c2a1ab57d28df16f3cf0b6cd72803ce638"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.915192 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5qtlg" podStartSLOduration=125.915168886 podStartE2EDuration="2m5.915168886s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:02.894319862 +0000 UTC m=+147.480910626" watchObservedRunningTime="2025-11-27 10:24:02.915168886 +0000 UTC m=+147.501759650" Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.951608 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-28rbb" event={"ID":"9b0fd0cb-13fa-4f87-bfc3-defe908884d4","Type":"ContainerStarted","Data":"317748e7ede618e76a34dba7994a1e8ede9d8e41a24c25b4b59fd518d3508e3b"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.954525 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" event={"ID":"4d69e866-000e-4748-9953-4d5f007f2669","Type":"ContainerStarted","Data":"ff5528ecbad477bebb1840320ebda37649f61c4e873f54415957ffe19fe239de"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.954546 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" event={"ID":"4d69e866-000e-4748-9953-4d5f007f2669","Type":"ContainerStarted","Data":"eec4c0580fc87d33500e73334003f7a135f9e47721e49314fd7f89b6e591ceb6"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.986492 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:02 crc kubenswrapper[4806]: E1127 10:24:02.987131 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.487071787 +0000 UTC m=+148.073662551 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.987876 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:02 crc kubenswrapper[4806]: E1127 10:24:02.988846 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.488833056 +0000 UTC m=+148.075424050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.989810 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" event={"ID":"7ce6a8cd-9071-4c23-a36f-9af525e7f04e","Type":"ContainerStarted","Data":"aa8ff45313b16d33e68b2e1c4a5aebbeaa37646cfb844b7cdd41b1e30d68e712"} Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.993876 4806 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7t7px container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 27 10:24:02 crc kubenswrapper[4806]: I1127 10:24:02.993913 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" podUID="7ce6a8cd-9071-4c23-a36f-9af525e7f04e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.001321 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" podStartSLOduration=126.001301386 podStartE2EDuration="2m6.001301386s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:02.994599638 +0000 UTC m=+147.581190402" watchObservedRunningTime="2025-11-27 10:24:03.001301386 +0000 UTC m=+147.587892150" Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.034284 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" event={"ID":"777fcbda-0e10-4a62-aa45-fa8670bae076","Type":"ContainerStarted","Data":"09011ebc44a3409ef4b40816272e01ba141031306e4552deee205718192e72c0"} Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.087508 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-gnrs8" podStartSLOduration=126.087483796 podStartE2EDuration="2m6.087483796s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:03.085804999 +0000 UTC m=+147.672395763" watchObservedRunningTime="2025-11-27 10:24:03.087483796 +0000 UTC m=+147.674074560" Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.090963 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:03 crc kubenswrapper[4806]: E1127 10:24:03.092034 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.592021973 +0000 UTC m=+148.178612737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.107466 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" event={"ID":"06f03ec3-796d-442f-8adc-93aa22466426","Type":"ContainerStarted","Data":"463f8cd933504792e79065603dc6d6976d0e8f84cb5d2739d7622a48621636d7"} Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.127858 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.172925 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:03 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:03 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:03 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.172996 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.194057 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:03 crc kubenswrapper[4806]: E1127 10:24:03.201497 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.701480915 +0000 UTC m=+148.288071679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.202799 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" podStartSLOduration=126.20278231099999 podStartE2EDuration="2m6.202782311s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:03.134480281 +0000 UTC m=+147.721071045" watchObservedRunningTime="2025-11-27 10:24:03.202782311 +0000 UTC m=+147.789373065" Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.246752 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rnhfh" podStartSLOduration=126.246734711 podStartE2EDuration="2m6.246734711s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:03.208744159 +0000 UTC m=+147.795334923" watchObservedRunningTime="2025-11-27 10:24:03.246734711 +0000 UTC m=+147.833325475" Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.274993 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxg4x" Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.293555 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9" podStartSLOduration=127.29352517 podStartE2EDuration="2m7.29352517s" podCreationTimestamp="2025-11-27 10:21:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:03.249305013 +0000 UTC m=+147.835895777" watchObservedRunningTime="2025-11-27 10:24:03.29352517 +0000 UTC m=+147.880115934" Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.295769 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:03 crc kubenswrapper[4806]: E1127 10:24:03.296173 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.796154724 +0000 UTC m=+148.382745488 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.296374 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:03 crc kubenswrapper[4806]: E1127 10:24:03.297077 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.797069599 +0000 UTC m=+148.383660363 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.398324 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:03 crc kubenswrapper[4806]: E1127 10:24:03.398551 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.898520897 +0000 UTC m=+148.485111661 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.398808 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:03 crc kubenswrapper[4806]: E1127 10:24:03.399138 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:03.899127314 +0000 UTC m=+148.485718128 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.479204 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fszts" podStartSLOduration=126.479186603 podStartE2EDuration="2m6.479186603s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:03.373813156 +0000 UTC m=+147.960403920" watchObservedRunningTime="2025-11-27 10:24:03.479186603 +0000 UTC m=+148.065777367" Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.479839 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nflhv" podStartSLOduration=126.479835941 podStartE2EDuration="2m6.479835941s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:03.478214006 +0000 UTC m=+148.064804770" watchObservedRunningTime="2025-11-27 10:24:03.479835941 +0000 UTC m=+148.066426705" Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.500121 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:03 crc kubenswrapper[4806]: E1127 10:24:03.500302 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:04.000276484 +0000 UTC m=+148.586867238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.500459 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:03 crc kubenswrapper[4806]: E1127 10:24:03.500817 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:04.000802518 +0000 UTC m=+148.587393352 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.601769 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:03 crc kubenswrapper[4806]: E1127 10:24:03.602257 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:04.102240776 +0000 UTC m=+148.688831550 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.703329 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:03 crc kubenswrapper[4806]: E1127 10:24:03.703612 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:04.203601071 +0000 UTC m=+148.790191835 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.725667 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.804802 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:03 crc kubenswrapper[4806]: E1127 10:24:03.805853 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:04.305834771 +0000 UTC m=+148.892425535 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.906979 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.907070 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:03 crc kubenswrapper[4806]: E1127 10:24:03.908190 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:04.408174734 +0000 UTC m=+148.994765498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:03 crc kubenswrapper[4806]: I1127 10:24:03.915163 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.008605 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:04 crc kubenswrapper[4806]: E1127 10:24:04.008790 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:04.508762678 +0000 UTC m=+149.095353442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.008854 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.008902 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.009065 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.009134 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:24:04 crc kubenswrapper[4806]: E1127 10:24:04.010578 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:04.510564198 +0000 UTC m=+149.097154962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.025189 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.032519 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.032648 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.039159 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.050903 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.055536 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.128125 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:04 crc kubenswrapper[4806]: E1127 10:24:04.129026 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:04.629000651 +0000 UTC m=+149.215591415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.176480 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:04 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:04 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:04 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.176542 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.200830 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-r7wlb" event={"ID":"e858c9d3-3f1c-47db-a0c8-f0efac1ed114","Type":"ContainerStarted","Data":"2613cea55c8426e264cf011e1862e552a7e1027fbe9b33e33ef83552ff5e9e78"} Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.212760 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z979k" event={"ID":"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43","Type":"ContainerStarted","Data":"0d6822859e3ec1d34f078b53953cca661e456beac0acb007dd7e86568fd816a6"} Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.230726 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:04 crc kubenswrapper[4806]: E1127 10:24:04.231225 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:04.731207951 +0000 UTC m=+149.317798715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.234143 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" event={"ID":"67a0b3a8-bfaa-48dc-8a19-d3b0ef450b06","Type":"ContainerStarted","Data":"6fe465153c06622a462f14c79650b05da2d530b25228930a1cff5d7afc495b7f"} Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.262514 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-r7wlb" podStartSLOduration=127.262486986 podStartE2EDuration="2m7.262486986s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:04.261351073 +0000 UTC m=+148.847941827" watchObservedRunningTime="2025-11-27 10:24:04.262486986 +0000 UTC m=+148.849077750" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.282047 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qgxx9" event={"ID":"a3ce016c-a670-46a3-9254-03fea2c6568b","Type":"ContainerStarted","Data":"b590245f8a9b1bec13aabceb5491549988a1f78a93c5e3d9bacb935ee6331639"} Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.320914 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lzwl4" event={"ID":"f9ef04e5-e987-4457-bd8b-d1bc01cb67ac","Type":"ContainerStarted","Data":"14aee04f4f8b9698a7c6489809a5c3d0c197e9eb2e9ef3b5ecb90ee92ed66f69"} Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.331476 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:04 crc kubenswrapper[4806]: E1127 10:24:04.332039 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:04.832009651 +0000 UTC m=+149.418600415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.333347 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:04 crc kubenswrapper[4806]: E1127 10:24:04.336087 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:04.836070574 +0000 UTC m=+149.422661338 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.339045 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zbhcm" event={"ID":"f6fc593a-b4b9-46ee-b12f-9ac4732dc69e","Type":"ContainerStarted","Data":"ea025d7215b8eba9c359b5a2eb0b6e3e1f5acb69a1c1862f19cf8975d20a03b4"} Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.339433 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-zbhcm" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.372407 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-prj6n" podStartSLOduration=127.37238399 podStartE2EDuration="2m7.37238399s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:04.318663657 +0000 UTC m=+148.905254421" watchObservedRunningTime="2025-11-27 10:24:04.37238399 +0000 UTC m=+148.958974754" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.396558 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" event={"ID":"a52b3a70-665d-49e5-a842-14ade4dcff1f","Type":"ContainerStarted","Data":"34eaa5f8414ff403ce2cef3f0ad220ffa0ff83d82ed36aaaa4dedb8b0f20d113"} Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.397544 4806 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-xvtnq container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.397609 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" podUID="3fb2578f-5e34-45b4-bcae-4798bfba073a" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.399440 4806 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7t7px container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.399517 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" podUID="7ce6a8cd-9071-4c23-a36f-9af525e7f04e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.431259 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-lzwl4" podStartSLOduration=127.431219355 podStartE2EDuration="2m7.431219355s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:04.374573231 +0000 UTC m=+148.961163995" watchObservedRunningTime="2025-11-27 10:24:04.431219355 +0000 UTC m=+149.017810119" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.434293 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:04 crc kubenswrapper[4806]: E1127 10:24:04.434865 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:04.934830106 +0000 UTC m=+149.521420870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.536556 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:04 crc kubenswrapper[4806]: E1127 10:24:04.540559 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:05.040545804 +0000 UTC m=+149.627136568 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.640452 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:04 crc kubenswrapper[4806]: E1127 10:24:04.641132 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:05.141113917 +0000 UTC m=+149.727704681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.683446 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-zbhcm" podStartSLOduration=10.68341632 podStartE2EDuration="10.68341632s" podCreationTimestamp="2025-11-27 10:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:04.433362056 +0000 UTC m=+149.019952820" watchObservedRunningTime="2025-11-27 10:24:04.68341632 +0000 UTC m=+149.270007084" Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.742505 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:04 crc kubenswrapper[4806]: E1127 10:24:04.743000 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:05.242978886 +0000 UTC m=+149.829569650 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.843815 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:04 crc kubenswrapper[4806]: E1127 10:24:04.844262 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:05.344224548 +0000 UTC m=+149.930815312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:04 crc kubenswrapper[4806]: I1127 10:24:04.945325 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:04 crc kubenswrapper[4806]: E1127 10:24:04.945711 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:05.445698868 +0000 UTC m=+150.032289632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.049460 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:05 crc kubenswrapper[4806]: E1127 10:24:05.050002 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:05.549980505 +0000 UTC m=+150.136571269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.151040 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:05 crc kubenswrapper[4806]: E1127 10:24:05.151655 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:05.651625218 +0000 UTC m=+150.238216162 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.182304 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:05 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:05 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:05 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.182359 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.256140 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:05 crc kubenswrapper[4806]: E1127 10:24:05.257466 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:05.757441098 +0000 UTC m=+150.344031872 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.359996 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:05 crc kubenswrapper[4806]: E1127 10:24:05.360348 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:05.860335987 +0000 UTC m=+150.446926751 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.398241 4806 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wsjh5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.398350 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" podUID="16a0adb3-9386-417e-8a34-bdf429462e2b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.422402 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z979k" event={"ID":"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43","Type":"ContainerStarted","Data":"0e94f93dc9b915350a8125dc72336aab270621c0c705ac3ad8d9c17c816d66bb"} Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.461146 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:05 crc kubenswrapper[4806]: E1127 10:24:05.461440 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:05.961396884 +0000 UTC m=+150.547987648 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.461805 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:05 crc kubenswrapper[4806]: E1127 10:24:05.464108 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:05.964080309 +0000 UTC m=+150.550671233 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.571748 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:05 crc kubenswrapper[4806]: E1127 10:24:05.572103 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:06.072079839 +0000 UTC m=+150.658670603 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.688128 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:05 crc kubenswrapper[4806]: E1127 10:24:05.688600 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:06.188581859 +0000 UTC m=+150.775172623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.789388 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:05 crc kubenswrapper[4806]: E1127 10:24:05.790280 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:06.290257433 +0000 UTC m=+150.876848197 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.812280 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" podStartSLOduration=129.812259008 podStartE2EDuration="2m9.812259008s" podCreationTimestamp="2025-11-27 10:21:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:04.69271572 +0000 UTC m=+149.279306484" watchObservedRunningTime="2025-11-27 10:24:05.812259008 +0000 UTC m=+150.398849772" Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.891562 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:05 crc kubenswrapper[4806]: E1127 10:24:05.892016 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:06.391994718 +0000 UTC m=+150.978585482 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:05 crc kubenswrapper[4806]: I1127 10:24:05.993804 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:05 crc kubenswrapper[4806]: E1127 10:24:05.994115 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:06.494101795 +0000 UTC m=+151.080692559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.095870 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:06 crc kubenswrapper[4806]: E1127 10:24:06.096287 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:06.596267413 +0000 UTC m=+151.182858177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.167604 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:06 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:06 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:06 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.167681 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.196850 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:06 crc kubenswrapper[4806]: E1127 10:24:06.197024 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:06.697001511 +0000 UTC m=+151.283592275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.197209 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:06 crc kubenswrapper[4806]: E1127 10:24:06.197503 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:06.697492114 +0000 UTC m=+151.284082878 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:06 crc kubenswrapper[4806]: W1127 10:24:06.227895 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-9977ff2480e5663e36db2f899c6642b9dc20a4a8944672e9b31ed3a2833ee9a5 WatchSource:0}: Error finding container 9977ff2480e5663e36db2f899c6642b9dc20a4a8944672e9b31ed3a2833ee9a5: Status 404 returned error can't find the container with id 9977ff2480e5663e36db2f899c6642b9dc20a4a8944672e9b31ed3a2833ee9a5 Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.297818 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:06 crc kubenswrapper[4806]: E1127 10:24:06.297961 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:06.797924094 +0000 UTC m=+151.384514858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.298433 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:06 crc kubenswrapper[4806]: E1127 10:24:06.306661 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:06.806626157 +0000 UTC m=+151.393216921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.402965 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:06 crc kubenswrapper[4806]: E1127 10:24:06.403341 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:06.903327053 +0000 UTC m=+151.489917817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.423646 4806 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wsjh5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.423710 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" podUID="16a0adb3-9386-417e-8a34-bdf429462e2b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.454425 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"9977ff2480e5663e36db2f899c6642b9dc20a4a8944672e9b31ed3a2833ee9a5"} Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.468342 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4e1bca5c5c4821c3e1db28deed9c5473da3505895ea8da9a8fbd71a7ba4ebfb0"} Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.477385 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"6e279b15386afb2575d72bde88fb34943feed07b7da47c505db6c7512b76d06f"} Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.492141 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z979k" event={"ID":"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43","Type":"ContainerStarted","Data":"79333a77df5bb07f1299fed1a85e9a5f9fcf65ddaa677ef82faa012289894620"} Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.505978 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:06 crc kubenswrapper[4806]: E1127 10:24:06.506351 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:07.006340205 +0000 UTC m=+151.592930969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.608038 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:06 crc kubenswrapper[4806]: E1127 10:24:06.608530 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:07.108506322 +0000 UTC m=+151.695097086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.709324 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jprw7"] Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.709646 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:06 crc kubenswrapper[4806]: E1127 10:24:06.709951 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:07.209938339 +0000 UTC m=+151.796529103 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.711222 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.714248 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.811257 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:06 crc kubenswrapper[4806]: E1127 10:24:06.811603 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:07.311531742 +0000 UTC m=+151.898122506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.811864 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.811906 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw5vm\" (UniqueName: \"kubernetes.io/projected/a3bc679e-e16d-47ce-9015-c110b1604ac1-kube-api-access-zw5vm\") pod \"community-operators-jprw7\" (UID: \"a3bc679e-e16d-47ce-9015-c110b1604ac1\") " pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.811930 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3bc679e-e16d-47ce-9015-c110b1604ac1-utilities\") pod \"community-operators-jprw7\" (UID: \"a3bc679e-e16d-47ce-9015-c110b1604ac1\") " pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.811955 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3bc679e-e16d-47ce-9015-c110b1604ac1-catalog-content\") pod \"community-operators-jprw7\" (UID: \"a3bc679e-e16d-47ce-9015-c110b1604ac1\") " pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:24:06 crc kubenswrapper[4806]: E1127 10:24:06.812528 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:07.312516949 +0000 UTC m=+151.899107893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.912983 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:06 crc kubenswrapper[4806]: E1127 10:24:06.913198 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:07.413156115 +0000 UTC m=+151.999746879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.913291 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw5vm\" (UniqueName: \"kubernetes.io/projected/a3bc679e-e16d-47ce-9015-c110b1604ac1-kube-api-access-zw5vm\") pod \"community-operators-jprw7\" (UID: \"a3bc679e-e16d-47ce-9015-c110b1604ac1\") " pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.913373 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3bc679e-e16d-47ce-9015-c110b1604ac1-utilities\") pod \"community-operators-jprw7\" (UID: \"a3bc679e-e16d-47ce-9015-c110b1604ac1\") " pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.913452 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3bc679e-e16d-47ce-9015-c110b1604ac1-catalog-content\") pod \"community-operators-jprw7\" (UID: \"a3bc679e-e16d-47ce-9015-c110b1604ac1\") " pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.913819 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3bc679e-e16d-47ce-9015-c110b1604ac1-utilities\") pod \"community-operators-jprw7\" (UID: \"a3bc679e-e16d-47ce-9015-c110b1604ac1\") " pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.913938 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3bc679e-e16d-47ce-9015-c110b1604ac1-catalog-content\") pod \"community-operators-jprw7\" (UID: \"a3bc679e-e16d-47ce-9015-c110b1604ac1\") " pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.938894 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.938970 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:24:06 crc kubenswrapper[4806]: I1127 10:24:06.958902 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw5vm\" (UniqueName: \"kubernetes.io/projected/a3bc679e-e16d-47ce-9015-c110b1604ac1-kube-api-access-zw5vm\") pod \"community-operators-jprw7\" (UID: \"a3bc679e-e16d-47ce-9015-c110b1604ac1\") " pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.015310 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:07 crc kubenswrapper[4806]: E1127 10:24:07.016002 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:07.515987131 +0000 UTC m=+152.102577895 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.028873 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.065537 4806 patch_prober.go:28] interesting pod/downloads-7954f5f757-9ppfs container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.065623 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9ppfs" podUID="52dd6d6d-3a15-4467-9d82-37fd2fa5794b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.066202 4806 patch_prober.go:28] interesting pod/downloads-7954f5f757-9ppfs container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.066255 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9ppfs" podUID="52dd6d6d-3a15-4467-9d82-37fd2fa5794b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.069706 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.070881 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.085014 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.097735 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-42h4s" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.116771 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:07 crc kubenswrapper[4806]: E1127 10:24:07.117661 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:07.617643705 +0000 UTC m=+152.204234469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.142179 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jprw7"] Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.146806 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9v4rb"] Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.148201 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.177438 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:07 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:07 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:07 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.177493 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.220276 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.222530 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3c3761a-7b4a-4053-bf59-0634f90d1e91-utilities\") pod \"community-operators-9v4rb\" (UID: \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\") " pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:24:07 crc kubenswrapper[4806]: E1127 10:24:07.222724 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:07.722671772 +0000 UTC m=+152.309262536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.222795 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3c3761a-7b4a-4053-bf59-0634f90d1e91-catalog-content\") pod \"community-operators-9v4rb\" (UID: \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\") " pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.223048 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ctb5\" (UniqueName: \"kubernetes.io/projected/c3c3761a-7b4a-4053-bf59-0634f90d1e91-kube-api-access-6ctb5\") pod \"community-operators-9v4rb\" (UID: \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\") " pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.234811 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9v4rb"] Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.310438 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wfw97"] Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.311880 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.327209 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.327514 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3c3761a-7b4a-4053-bf59-0634f90d1e91-utilities\") pod \"community-operators-9v4rb\" (UID: \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\") " pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.327550 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3c3761a-7b4a-4053-bf59-0634f90d1e91-catalog-content\") pod \"community-operators-9v4rb\" (UID: \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\") " pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.327614 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ctb5\" (UniqueName: \"kubernetes.io/projected/c3c3761a-7b4a-4053-bf59-0634f90d1e91-kube-api-access-6ctb5\") pod \"community-operators-9v4rb\" (UID: \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\") " pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.328382 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3c3761a-7b4a-4053-bf59-0634f90d1e91-utilities\") pod \"community-operators-9v4rb\" (UID: \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\") " pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:24:07 crc kubenswrapper[4806]: E1127 10:24:07.328599 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:07.828570925 +0000 UTC m=+152.415161689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.330989 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3c3761a-7b4a-4053-bf59-0634f90d1e91-catalog-content\") pod \"community-operators-9v4rb\" (UID: \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\") " pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.372455 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.405329 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wfw97"] Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.411794 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.411934 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.413677 4806 patch_prober.go:28] interesting pod/console-f9d7485db-6fqzk container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.413760 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6fqzk" podUID="97042267-61ca-4864-a50a-60587abc7619" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.429269 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.429329 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-catalog-content\") pod \"certified-operators-wfw97\" (UID: \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\") " pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.429393 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-utilities\") pod \"certified-operators-wfw97\" (UID: \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\") " pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.429431 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf426\" (UniqueName: \"kubernetes.io/projected/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-kube-api-access-tf426\") pod \"certified-operators-wfw97\" (UID: \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\") " pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:24:07 crc kubenswrapper[4806]: E1127 10:24:07.429854 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:07.929837478 +0000 UTC m=+152.516428242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.499688 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ee95ee28de0f25352f0d13f0d070fab81f8344ab0bc75fc6fd5bcbfce8feba7a"} Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.503860 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z979k" event={"ID":"d3b66f29-69bb-4b08-b3bd-ce924dbe7d43","Type":"ContainerStarted","Data":"df2020a8804d77ed3d18a4d90f83ec719973e2f834bf565f1796118d0f975167"} Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.523167 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ctb5\" (UniqueName: \"kubernetes.io/projected/c3c3761a-7b4a-4053-bf59-0634f90d1e91-kube-api-access-6ctb5\") pod \"community-operators-9v4rb\" (UID: \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\") " pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.530504 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.530825 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-catalog-content\") pod \"certified-operators-wfw97\" (UID: \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\") " pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.531018 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-utilities\") pod \"certified-operators-wfw97\" (UID: \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\") " pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.531084 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf426\" (UniqueName: \"kubernetes.io/projected/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-kube-api-access-tf426\") pod \"certified-operators-wfw97\" (UID: \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\") " pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.533977 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ea1eaed384b655fe883db5f901332aa2943c63ed4ac45a39d646a8c117b242fc"} Nov 27 10:24:07 crc kubenswrapper[4806]: E1127 10:24:07.538284 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:08.038253011 +0000 UTC m=+152.624843775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.560769 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"75e33408fef3759d3f18c8d5943a838c70795a550e7fb42e6fccf1ea3c2ca4ba"} Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.560872 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.563891 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-utilities\") pod \"certified-operators-wfw97\" (UID: \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\") " pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.564141 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-catalog-content\") pod \"certified-operators-wfw97\" (UID: \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\") " pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.613460 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bsqsc" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.635459 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.638048 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf426\" (UniqueName: \"kubernetes.io/projected/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-kube-api-access-tf426\") pod \"certified-operators-wfw97\" (UID: \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\") " pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:24:07 crc kubenswrapper[4806]: E1127 10:24:07.639537 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:08.139521364 +0000 UTC m=+152.726112128 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.651599 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2cthc"] Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.653696 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.662333 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2cthc"] Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.752849 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.753298 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mz5t\" (UniqueName: \"kubernetes.io/projected/ae461a45-1190-46fc-acf5-8954b951b25e-kube-api-access-7mz5t\") pod \"certified-operators-2cthc\" (UID: \"ae461a45-1190-46fc-acf5-8954b951b25e\") " pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.753349 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae461a45-1190-46fc-acf5-8954b951b25e-catalog-content\") pod \"certified-operators-2cthc\" (UID: \"ae461a45-1190-46fc-acf5-8954b951b25e\") " pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.753367 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae461a45-1190-46fc-acf5-8954b951b25e-utilities\") pod \"certified-operators-2cthc\" (UID: \"ae461a45-1190-46fc-acf5-8954b951b25e\") " pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:24:07 crc kubenswrapper[4806]: E1127 10:24:07.753486 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:08.253471952 +0000 UTC m=+152.840062716 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.776708 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.857430 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae461a45-1190-46fc-acf5-8954b951b25e-catalog-content\") pod \"certified-operators-2cthc\" (UID: \"ae461a45-1190-46fc-acf5-8954b951b25e\") " pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.857810 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae461a45-1190-46fc-acf5-8954b951b25e-utilities\") pod \"certified-operators-2cthc\" (UID: \"ae461a45-1190-46fc-acf5-8954b951b25e\") " pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.857968 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.858072 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mz5t\" (UniqueName: \"kubernetes.io/projected/ae461a45-1190-46fc-acf5-8954b951b25e-kube-api-access-7mz5t\") pod \"certified-operators-2cthc\" (UID: \"ae461a45-1190-46fc-acf5-8954b951b25e\") " pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:24:07 crc kubenswrapper[4806]: E1127 10:24:07.859078 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:08.359061686 +0000 UTC m=+152.945652450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.859173 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae461a45-1190-46fc-acf5-8954b951b25e-utilities\") pod \"certified-operators-2cthc\" (UID: \"ae461a45-1190-46fc-acf5-8954b951b25e\") " pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.859665 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae461a45-1190-46fc-acf5-8954b951b25e-catalog-content\") pod \"certified-operators-2cthc\" (UID: \"ae461a45-1190-46fc-acf5-8954b951b25e\") " pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.937179 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.961837 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:07 crc kubenswrapper[4806]: E1127 10:24:07.962282 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:08.462257181 +0000 UTC m=+153.048847945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:07 crc kubenswrapper[4806]: I1127 10:24:07.969126 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mz5t\" (UniqueName: \"kubernetes.io/projected/ae461a45-1190-46fc-acf5-8954b951b25e-kube-api-access-7mz5t\") pod \"certified-operators-2cthc\" (UID: \"ae461a45-1190-46fc-acf5-8954b951b25e\") " pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.038062 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.039091 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xvtnq" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.063075 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:08 crc kubenswrapper[4806]: E1127 10:24:08.064112 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:08.56407877 +0000 UTC m=+153.150669534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.100420 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-z979k" podStartSLOduration=14.100386446 podStartE2EDuration="14.100386446s" podCreationTimestamp="2025-11-27 10:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:08.015679157 +0000 UTC m=+152.602269921" watchObservedRunningTime="2025-11-27 10:24:08.100386446 +0000 UTC m=+152.686977210" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.158934 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.163269 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.163548 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:08 crc kubenswrapper[4806]: E1127 10:24:08.163768 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:08.663733378 +0000 UTC m=+153.250324142 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.164038 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:08 crc kubenswrapper[4806]: E1127 10:24:08.164501 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:08.66449191 +0000 UTC m=+153.251082674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.179309 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:08 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:08 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:08 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.179846 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.233441 4806 patch_prober.go:28] interesting pod/apiserver-76f77b778f-8jh8g container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 27 10:24:08 crc kubenswrapper[4806]: [+]log ok Nov 27 10:24:08 crc kubenswrapper[4806]: [+]etcd ok Nov 27 10:24:08 crc kubenswrapper[4806]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 27 10:24:08 crc kubenswrapper[4806]: [+]poststarthook/generic-apiserver-start-informers ok Nov 27 10:24:08 crc kubenswrapper[4806]: [+]poststarthook/max-in-flight-filter ok Nov 27 10:24:08 crc kubenswrapper[4806]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 27 10:24:08 crc kubenswrapper[4806]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 27 10:24:08 crc kubenswrapper[4806]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 27 10:24:08 crc kubenswrapper[4806]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 27 10:24:08 crc kubenswrapper[4806]: [+]poststarthook/project.openshift.io-projectcache ok Nov 27 10:24:08 crc kubenswrapper[4806]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 27 10:24:08 crc kubenswrapper[4806]: [+]poststarthook/openshift.io-startinformers ok Nov 27 10:24:08 crc kubenswrapper[4806]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 27 10:24:08 crc kubenswrapper[4806]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 27 10:24:08 crc kubenswrapper[4806]: livez check failed Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.233557 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" podUID="a52b3a70-665d-49e5-a842-14ade4dcff1f" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.237680 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wsjh5" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.271211 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:08 crc kubenswrapper[4806]: E1127 10:24:08.272092 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:08.772072018 +0000 UTC m=+153.358662782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.373541 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:08 crc kubenswrapper[4806]: E1127 10:24:08.373952 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:08.873920067 +0000 UTC m=+153.460510831 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.401036 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jprw7"] Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.443017 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.443812 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.470690 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.473529 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.475290 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.475938 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66dfffe4-4026-409f-a8a2-bc8d23c55ba1-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"66dfffe4-4026-409f-a8a2-bc8d23c55ba1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.476290 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66dfffe4-4026-409f-a8a2-bc8d23c55ba1-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"66dfffe4-4026-409f-a8a2-bc8d23c55ba1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 10:24:08 crc kubenswrapper[4806]: E1127 10:24:08.476609 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:08.976578749 +0000 UTC m=+153.563169513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.532669 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.579618 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66dfffe4-4026-409f-a8a2-bc8d23c55ba1-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"66dfffe4-4026-409f-a8a2-bc8d23c55ba1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.579764 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66dfffe4-4026-409f-a8a2-bc8d23c55ba1-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"66dfffe4-4026-409f-a8a2-bc8d23c55ba1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.579834 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.580253 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66dfffe4-4026-409f-a8a2-bc8d23c55ba1-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"66dfffe4-4026-409f-a8a2-bc8d23c55ba1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.580328 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jprw7" event={"ID":"a3bc679e-e16d-47ce-9015-c110b1604ac1","Type":"ContainerStarted","Data":"8d5ee725a6b61939fce7eb307e1e96757a3d4f8c5cff037c49250d44d8ac6fb8"} Nov 27 10:24:08 crc kubenswrapper[4806]: E1127 10:24:08.588779 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:09.088736746 +0000 UTC m=+153.675327510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.635528 4806 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.685380 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:08 crc kubenswrapper[4806]: E1127 10:24:08.688055 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:09.188031795 +0000 UTC m=+153.774622559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.691434 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66dfffe4-4026-409f-a8a2-bc8d23c55ba1-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"66dfffe4-4026-409f-a8a2-bc8d23c55ba1\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.792537 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:08 crc kubenswrapper[4806]: E1127 10:24:08.793038 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:09.293017731 +0000 UTC m=+153.879608495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.810532 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.894821 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:08 crc kubenswrapper[4806]: E1127 10:24:08.895543 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:09.395524989 +0000 UTC m=+153.982115753 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:08 crc kubenswrapper[4806]: I1127 10:24:08.998164 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:08 crc kubenswrapper[4806]: E1127 10:24:08.998575 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:09.498563571 +0000 UTC m=+154.085154335 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.099634 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:09 crc kubenswrapper[4806]: E1127 10:24:09.100258 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:09.600209585 +0000 UTC m=+154.186800349 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.100482 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:09 crc kubenswrapper[4806]: E1127 10:24:09.100915 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:09.600896574 +0000 UTC m=+154.187487338 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.175560 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:09 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:09 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:09 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.175638 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.202012 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:09 crc kubenswrapper[4806]: E1127 10:24:09.202434 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:09.702419513 +0000 UTC m=+154.289010277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.300290 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gp6rt"] Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.301343 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.305134 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:09 crc kubenswrapper[4806]: E1127 10:24:09.305632 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:09.80561673 +0000 UTC m=+154.392207504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.324075 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.406538 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.406828 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3209cb7e-221c-4c6a-9864-8101e8171078-catalog-content\") pod \"redhat-marketplace-gp6rt\" (UID: \"3209cb7e-221c-4c6a-9864-8101e8171078\") " pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.406890 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3209cb7e-221c-4c6a-9864-8101e8171078-utilities\") pod \"redhat-marketplace-gp6rt\" (UID: \"3209cb7e-221c-4c6a-9864-8101e8171078\") " pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.406925 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5qgh\" (UniqueName: \"kubernetes.io/projected/3209cb7e-221c-4c6a-9864-8101e8171078-kube-api-access-w5qgh\") pod \"redhat-marketplace-gp6rt\" (UID: \"3209cb7e-221c-4c6a-9864-8101e8171078\") " pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:24:09 crc kubenswrapper[4806]: E1127 10:24:09.407077 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:09.907062538 +0000 UTC m=+154.493653302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.470952 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp6rt"] Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.484116 4806 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-27T10:24:08.635580237Z","Handler":null,"Name":""} Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.508630 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3209cb7e-221c-4c6a-9864-8101e8171078-catalog-content\") pod \"redhat-marketplace-gp6rt\" (UID: \"3209cb7e-221c-4c6a-9864-8101e8171078\") " pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.508696 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.508734 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3209cb7e-221c-4c6a-9864-8101e8171078-utilities\") pod \"redhat-marketplace-gp6rt\" (UID: \"3209cb7e-221c-4c6a-9864-8101e8171078\") " pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.508782 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5qgh\" (UniqueName: \"kubernetes.io/projected/3209cb7e-221c-4c6a-9864-8101e8171078-kube-api-access-w5qgh\") pod \"redhat-marketplace-gp6rt\" (UID: \"3209cb7e-221c-4c6a-9864-8101e8171078\") " pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:24:09 crc kubenswrapper[4806]: E1127 10:24:09.509082 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:10.009062952 +0000 UTC m=+154.595653716 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.509372 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3209cb7e-221c-4c6a-9864-8101e8171078-catalog-content\") pod \"redhat-marketplace-gp6rt\" (UID: \"3209cb7e-221c-4c6a-9864-8101e8171078\") " pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.509473 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3209cb7e-221c-4c6a-9864-8101e8171078-utilities\") pod \"redhat-marketplace-gp6rt\" (UID: \"3209cb7e-221c-4c6a-9864-8101e8171078\") " pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.550486 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5qgh\" (UniqueName: \"kubernetes.io/projected/3209cb7e-221c-4c6a-9864-8101e8171078-kube-api-access-w5qgh\") pod \"redhat-marketplace-gp6rt\" (UID: \"3209cb7e-221c-4c6a-9864-8101e8171078\") " pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.588684 4806 generic.go:334] "Generic (PLEG): container finished" podID="a3bc679e-e16d-47ce-9015-c110b1604ac1" containerID="9802e49e13468f6bc17d0afb1d7c68adc57f35aa536d4437bdf25bec0929fadf" exitCode=0 Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.588818 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jprw7" event={"ID":"a3bc679e-e16d-47ce-9015-c110b1604ac1","Type":"ContainerDied","Data":"9802e49e13468f6bc17d0afb1d7c68adc57f35aa536d4437bdf25bec0929fadf"} Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.591723 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.612937 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:09 crc kubenswrapper[4806]: E1127 10:24:09.613419 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:10.113372309 +0000 UTC m=+154.699963073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.656636 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.660029 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2cthc"] Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.677063 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wfw97"] Nov 27 10:24:09 crc kubenswrapper[4806]: W1127 10:24:09.732491 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae461a45_1190_46fc_acf5_8954b951b25e.slice/crio-f50a9bb1d7b1ceb7a6370bfff88a3b3d8e38d7328e2c39a7fbaf697bcc3ad46b WatchSource:0}: Error finding container f50a9bb1d7b1ceb7a6370bfff88a3b3d8e38d7328e2c39a7fbaf697bcc3ad46b: Status 404 returned error can't find the container with id f50a9bb1d7b1ceb7a6370bfff88a3b3d8e38d7328e2c39a7fbaf697bcc3ad46b Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.742054 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:09 crc kubenswrapper[4806]: E1127 10:24:09.742535 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-27 10:24:10.242518822 +0000 UTC m=+154.829109576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kmbdt" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.767025 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nxtlg"] Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.768398 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:24:09 crc kubenswrapper[4806]: W1127 10:24:09.788666 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba6aacd9_e39c_4a2c_bcfc_f5d1c252bdb2.slice/crio-e10cafd34ba5d265ee51dcdc7936eb1c1779ce58da22bde0c85d96551d80191b WatchSource:0}: Error finding container e10cafd34ba5d265ee51dcdc7936eb1c1779ce58da22bde0c85d96551d80191b: Status 404 returned error can't find the container with id e10cafd34ba5d265ee51dcdc7936eb1c1779ce58da22bde0c85d96551d80191b Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.800705 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxtlg"] Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.842942 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.843265 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72810aad-3a98-4ed6-bf65-6434e72770fb-utilities\") pod \"redhat-marketplace-nxtlg\" (UID: \"72810aad-3a98-4ed6-bf65-6434e72770fb\") " pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.843337 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72810aad-3a98-4ed6-bf65-6434e72770fb-catalog-content\") pod \"redhat-marketplace-nxtlg\" (UID: \"72810aad-3a98-4ed6-bf65-6434e72770fb\") " pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.843360 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjqq7\" (UniqueName: \"kubernetes.io/projected/72810aad-3a98-4ed6-bf65-6434e72770fb-kube-api-access-mjqq7\") pod \"redhat-marketplace-nxtlg\" (UID: \"72810aad-3a98-4ed6-bf65-6434e72770fb\") " pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:24:09 crc kubenswrapper[4806]: E1127 10:24:09.843467 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-27 10:24:10.343452716 +0000 UTC m=+154.930043480 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.924197 4806 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.924246 4806 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.945074 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.955692 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72810aad-3a98-4ed6-bf65-6434e72770fb-catalog-content\") pod \"redhat-marketplace-nxtlg\" (UID: \"72810aad-3a98-4ed6-bf65-6434e72770fb\") " pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.957247 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjqq7\" (UniqueName: \"kubernetes.io/projected/72810aad-3a98-4ed6-bf65-6434e72770fb-kube-api-access-mjqq7\") pod \"redhat-marketplace-nxtlg\" (UID: \"72810aad-3a98-4ed6-bf65-6434e72770fb\") " pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.957903 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72810aad-3a98-4ed6-bf65-6434e72770fb-utilities\") pod \"redhat-marketplace-nxtlg\" (UID: \"72810aad-3a98-4ed6-bf65-6434e72770fb\") " pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.956924 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72810aad-3a98-4ed6-bf65-6434e72770fb-catalog-content\") pod \"redhat-marketplace-nxtlg\" (UID: \"72810aad-3a98-4ed6-bf65-6434e72770fb\") " pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:24:09 crc kubenswrapper[4806]: I1127 10:24:09.958488 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72810aad-3a98-4ed6-bf65-6434e72770fb-utilities\") pod \"redhat-marketplace-nxtlg\" (UID: \"72810aad-3a98-4ed6-bf65-6434e72770fb\") " pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.065213 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjqq7\" (UniqueName: \"kubernetes.io/projected/72810aad-3a98-4ed6-bf65-6434e72770fb-kube-api-access-mjqq7\") pod \"redhat-marketplace-nxtlg\" (UID: \"72810aad-3a98-4ed6-bf65-6434e72770fb\") " pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.084601 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9v4rb"] Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.110887 4806 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.110960 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.137172 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.169535 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nmrrm"] Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.169593 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:10 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:10 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:10 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.169643 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.170765 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.176670 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.199994 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nmrrm"] Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.287469 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/293a1c8c-8131-462c-9dc4-4a8c516478c5-catalog-content\") pod \"redhat-operators-nmrrm\" (UID: \"293a1c8c-8131-462c-9dc4-4a8c516478c5\") " pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.287968 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/293a1c8c-8131-462c-9dc4-4a8c516478c5-utilities\") pod \"redhat-operators-nmrrm\" (UID: \"293a1c8c-8131-462c-9dc4-4a8c516478c5\") " pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.288195 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbvv8\" (UniqueName: \"kubernetes.io/projected/293a1c8c-8131-462c-9dc4-4a8c516478c5-kube-api-access-rbvv8\") pod \"redhat-operators-nmrrm\" (UID: \"293a1c8c-8131-462c-9dc4-4a8c516478c5\") " pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.316755 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l4x5q"] Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.318073 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.396584 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/293a1c8c-8131-462c-9dc4-4a8c516478c5-utilities\") pod \"redhat-operators-nmrrm\" (UID: \"293a1c8c-8131-462c-9dc4-4a8c516478c5\") " pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.397168 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbvv8\" (UniqueName: \"kubernetes.io/projected/293a1c8c-8131-462c-9dc4-4a8c516478c5-kube-api-access-rbvv8\") pod \"redhat-operators-nmrrm\" (UID: \"293a1c8c-8131-462c-9dc4-4a8c516478c5\") " pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.397408 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/293a1c8c-8131-462c-9dc4-4a8c516478c5-catalog-content\") pod \"redhat-operators-nmrrm\" (UID: \"293a1c8c-8131-462c-9dc4-4a8c516478c5\") " pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.403075 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/293a1c8c-8131-462c-9dc4-4a8c516478c5-catalog-content\") pod \"redhat-operators-nmrrm\" (UID: \"293a1c8c-8131-462c-9dc4-4a8c516478c5\") " pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.403287 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/293a1c8c-8131-462c-9dc4-4a8c516478c5-utilities\") pod \"redhat-operators-nmrrm\" (UID: \"293a1c8c-8131-462c-9dc4-4a8c516478c5\") " pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.437462 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.446272 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l4x5q"] Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.502434 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cb0392a-70a2-48c0-a69e-4421dbefe86e-catalog-content\") pod \"redhat-operators-l4x5q\" (UID: \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\") " pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.502883 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cb0392a-70a2-48c0-a69e-4421dbefe86e-utilities\") pod \"redhat-operators-l4x5q\" (UID: \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\") " pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.503058 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z57v\" (UniqueName: \"kubernetes.io/projected/8cb0392a-70a2-48c0-a69e-4421dbefe86e-kube-api-access-8z57v\") pod \"redhat-operators-l4x5q\" (UID: \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\") " pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.517279 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbvv8\" (UniqueName: \"kubernetes.io/projected/293a1c8c-8131-462c-9dc4-4a8c516478c5-kube-api-access-rbvv8\") pod \"redhat-operators-nmrrm\" (UID: \"293a1c8c-8131-462c-9dc4-4a8c516478c5\") " pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.527397 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kmbdt\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.604948 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.605679 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z57v\" (UniqueName: \"kubernetes.io/projected/8cb0392a-70a2-48c0-a69e-4421dbefe86e-kube-api-access-8z57v\") pod \"redhat-operators-l4x5q\" (UID: \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\") " pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.605765 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cb0392a-70a2-48c0-a69e-4421dbefe86e-catalog-content\") pod \"redhat-operators-l4x5q\" (UID: \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\") " pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.605836 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cb0392a-70a2-48c0-a69e-4421dbefe86e-utilities\") pod \"redhat-operators-l4x5q\" (UID: \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\") " pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.606394 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cb0392a-70a2-48c0-a69e-4421dbefe86e-utilities\") pod \"redhat-operators-l4x5q\" (UID: \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\") " pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.607016 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cb0392a-70a2-48c0-a69e-4421dbefe86e-catalog-content\") pod \"redhat-operators-l4x5q\" (UID: \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\") " pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.631812 4806 generic.go:334] "Generic (PLEG): container finished" podID="ae461a45-1190-46fc-acf5-8954b951b25e" containerID="31f320c341b64c5a2dadedd741fe3c4ad5719b3cb977c1ddb38df188b200a04b" exitCode=0 Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.631922 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cthc" event={"ID":"ae461a45-1190-46fc-acf5-8954b951b25e","Type":"ContainerDied","Data":"31f320c341b64c5a2dadedd741fe3c4ad5719b3cb977c1ddb38df188b200a04b"} Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.631968 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cthc" event={"ID":"ae461a45-1190-46fc-acf5-8954b951b25e","Type":"ContainerStarted","Data":"f50a9bb1d7b1ceb7a6370bfff88a3b3d8e38d7328e2c39a7fbaf697bcc3ad46b"} Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.688650 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.698726 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z57v\" (UniqueName: \"kubernetes.io/projected/8cb0392a-70a2-48c0-a69e-4421dbefe86e-kube-api-access-8z57v\") pod \"redhat-operators-l4x5q\" (UID: \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\") " pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.729608 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfw97" event={"ID":"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2","Type":"ContainerStarted","Data":"91e31ee3fcd48668d118ff226542405e05bc0e7a2ba717204372e4d83edd9a0b"} Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.729855 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfw97" event={"ID":"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2","Type":"ContainerStarted","Data":"e10cafd34ba5d265ee51dcdc7936eb1c1779ce58da22bde0c85d96551d80191b"} Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.775660 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.789952 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"66dfffe4-4026-409f-a8a2-bc8d23c55ba1","Type":"ContainerStarted","Data":"fd68911ac3e6e9f4777c617d892b72e20c844327de1908f811db350e87fa80fc"} Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.790496 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.830703 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9v4rb" event={"ID":"c3c3761a-7b4a-4053-bf59-0634f90d1e91","Type":"ContainerStarted","Data":"7b8fd3fa27aa38822799a65bae0a1507d81a8aa5ad734e4b97c5ff52228fdbfe"} Nov 27 10:24:10 crc kubenswrapper[4806]: I1127 10:24:10.831255 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.183508 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:11 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:11 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:11 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.183564 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.230737 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.242304 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.266726 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.271377 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.285219 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.408606 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.458560 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b373728-67ae-4cf8-bd5a-0678e7123d46-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3b373728-67ae-4cf8-bd5a-0678e7123d46\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.459075 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b373728-67ae-4cf8-bd5a-0678e7123d46-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3b373728-67ae-4cf8-bd5a-0678e7123d46\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.460574 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxtlg"] Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.478253 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp6rt"] Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.560447 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b373728-67ae-4cf8-bd5a-0678e7123d46-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3b373728-67ae-4cf8-bd5a-0678e7123d46\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.560627 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b373728-67ae-4cf8-bd5a-0678e7123d46-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3b373728-67ae-4cf8-bd5a-0678e7123d46\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.561417 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b373728-67ae-4cf8-bd5a-0678e7123d46-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3b373728-67ae-4cf8-bd5a-0678e7123d46\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.619288 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b373728-67ae-4cf8-bd5a-0678e7123d46-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3b373728-67ae-4cf8-bd5a-0678e7123d46\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.671749 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.809129 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nmrrm"] Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.916457 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxtlg" event={"ID":"72810aad-3a98-4ed6-bf65-6434e72770fb","Type":"ContainerStarted","Data":"b79e6ab9132478dba2aec7a03522e189d5c0ecd284beb2c974db43a639bbdc5c"} Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.932119 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"66dfffe4-4026-409f-a8a2-bc8d23c55ba1","Type":"ContainerStarted","Data":"df1c059553679b4d048112a22be3c985937acc86017ddebad0d9739c7e0a4cc2"} Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.949729 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9v4rb" event={"ID":"c3c3761a-7b4a-4053-bf59-0634f90d1e91","Type":"ContainerDied","Data":"405bca87ed37359429e68438556faa68669bfcfdcdf9b7d05474938733777905"} Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.950476 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.962219 4806 generic.go:334] "Generic (PLEG): container finished" podID="c3c3761a-7b4a-4053-bf59-0634f90d1e91" containerID="405bca87ed37359429e68438556faa68669bfcfdcdf9b7d05474938733777905" exitCode=0 Nov 27 10:24:11 crc kubenswrapper[4806]: I1127 10:24:11.976781 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-8jh8g" Nov 27 10:24:12 crc kubenswrapper[4806]: I1127 10:24:12.002546 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp6rt" event={"ID":"3209cb7e-221c-4c6a-9864-8101e8171078","Type":"ContainerStarted","Data":"e266031e45ae8fdebdbafdf907c955d8a6fed6f604c3aab0e667bcbe7db602ff"} Nov 27 10:24:12 crc kubenswrapper[4806]: W1127 10:24:12.012100 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod293a1c8c_8131_462c_9dc4_4a8c516478c5.slice/crio-00034d52ab8eb1c2f52a74cc1ac329e234ca8cf520ec193d6dbbff536036f3f7 WatchSource:0}: Error finding container 00034d52ab8eb1c2f52a74cc1ac329e234ca8cf520ec193d6dbbff536036f3f7: Status 404 returned error can't find the container with id 00034d52ab8eb1c2f52a74cc1ac329e234ca8cf520ec193d6dbbff536036f3f7 Nov 27 10:24:12 crc kubenswrapper[4806]: I1127 10:24:12.043020 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.043003104 podStartE2EDuration="4.043003104s" podCreationTimestamp="2025-11-27 10:24:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:11.996277487 +0000 UTC m=+156.582868251" watchObservedRunningTime="2025-11-27 10:24:12.043003104 +0000 UTC m=+156.629593868" Nov 27 10:24:12 crc kubenswrapper[4806]: I1127 10:24:12.069899 4806 generic.go:334] "Generic (PLEG): container finished" podID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" containerID="91e31ee3fcd48668d118ff226542405e05bc0e7a2ba717204372e4d83edd9a0b" exitCode=0 Nov 27 10:24:12 crc kubenswrapper[4806]: I1127 10:24:12.069959 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfw97" event={"ID":"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2","Type":"ContainerDied","Data":"91e31ee3fcd48668d118ff226542405e05bc0e7a2ba717204372e4d83edd9a0b"} Nov 27 10:24:12 crc kubenswrapper[4806]: I1127 10:24:12.167654 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:12 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:12 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:12 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:12 crc kubenswrapper[4806]: I1127 10:24:12.167721 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:12 crc kubenswrapper[4806]: I1127 10:24:12.220257 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 27 10:24:12 crc kubenswrapper[4806]: I1127 10:24:12.220903 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kmbdt"] Nov 27 10:24:12 crc kubenswrapper[4806]: I1127 10:24:12.324946 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l4x5q"] Nov 27 10:24:12 crc kubenswrapper[4806]: W1127 10:24:12.403637 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cb0392a_70a2_48c0_a69e_4421dbefe86e.slice/crio-73ea34e538f10958ca6c07b967476ab1f12b959729382fa29bf8ec7de19ccfc3 WatchSource:0}: Error finding container 73ea34e538f10958ca6c07b967476ab1f12b959729382fa29bf8ec7de19ccfc3: Status 404 returned error can't find the container with id 73ea34e538f10958ca6c07b967476ab1f12b959729382fa29bf8ec7de19ccfc3 Nov 27 10:24:12 crc kubenswrapper[4806]: I1127 10:24:12.569001 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 27 10:24:12 crc kubenswrapper[4806]: I1127 10:24:12.964283 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-zbhcm" Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.085038 4806 generic.go:334] "Generic (PLEG): container finished" podID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" containerID="50c6ff7f5250333d4f238fb01996502c749316c4d66d04c0151d9903dbd3cb5a" exitCode=0 Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.085124 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4x5q" event={"ID":"8cb0392a-70a2-48c0-a69e-4421dbefe86e","Type":"ContainerDied","Data":"50c6ff7f5250333d4f238fb01996502c749316c4d66d04c0151d9903dbd3cb5a"} Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.085196 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4x5q" event={"ID":"8cb0392a-70a2-48c0-a69e-4421dbefe86e","Type":"ContainerStarted","Data":"73ea34e538f10958ca6c07b967476ab1f12b959729382fa29bf8ec7de19ccfc3"} Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.092437 4806 generic.go:334] "Generic (PLEG): container finished" podID="3209cb7e-221c-4c6a-9864-8101e8171078" containerID="485f96fbbca01e2926f50422a4174a0874318ee9463dbbe63675bc94bf86d2e9" exitCode=0 Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.092642 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp6rt" event={"ID":"3209cb7e-221c-4c6a-9864-8101e8171078","Type":"ContainerDied","Data":"485f96fbbca01e2926f50422a4174a0874318ee9463dbbe63675bc94bf86d2e9"} Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.107368 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3b373728-67ae-4cf8-bd5a-0678e7123d46","Type":"ContainerStarted","Data":"1f6bc37a304a1f08ed124f8ffc6a45924624dbecad6144ebe6a7dfb4bcf7b72f"} Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.126802 4806 generic.go:334] "Generic (PLEG): container finished" podID="293a1c8c-8131-462c-9dc4-4a8c516478c5" containerID="f8db139004fbced6a3df6579096a70ac39008713c096b920e2cdf3aefcf44f49" exitCode=0 Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.127119 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nmrrm" event={"ID":"293a1c8c-8131-462c-9dc4-4a8c516478c5","Type":"ContainerDied","Data":"f8db139004fbced6a3df6579096a70ac39008713c096b920e2cdf3aefcf44f49"} Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.127492 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nmrrm" event={"ID":"293a1c8c-8131-462c-9dc4-4a8c516478c5","Type":"ContainerStarted","Data":"00034d52ab8eb1c2f52a74cc1ac329e234ca8cf520ec193d6dbbff536036f3f7"} Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.134142 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" event={"ID":"e2a6f027-898d-4a60-aa19-00b6f54d2aac","Type":"ContainerStarted","Data":"f830241c193cfa12475072329c53166088006fa0850175b330859d3b71bf9ead"} Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.134175 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" event={"ID":"e2a6f027-898d-4a60-aa19-00b6f54d2aac","Type":"ContainerStarted","Data":"27b5372bda575891bd1567d47d1f0de3acc3b266792acb3269aba0ab7af15df2"} Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.134429 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.137043 4806 generic.go:334] "Generic (PLEG): container finished" podID="66dfffe4-4026-409f-a8a2-bc8d23c55ba1" containerID="df1c059553679b4d048112a22be3c985937acc86017ddebad0d9739c7e0a4cc2" exitCode=0 Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.137138 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"66dfffe4-4026-409f-a8a2-bc8d23c55ba1","Type":"ContainerDied","Data":"df1c059553679b4d048112a22be3c985937acc86017ddebad0d9739c7e0a4cc2"} Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.139998 4806 generic.go:334] "Generic (PLEG): container finished" podID="72810aad-3a98-4ed6-bf65-6434e72770fb" containerID="6477e21483913b695296563b76f5d9077c9069c21e5ffb88e8f2936638243d2a" exitCode=0 Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.140882 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxtlg" event={"ID":"72810aad-3a98-4ed6-bf65-6434e72770fb","Type":"ContainerDied","Data":"6477e21483913b695296563b76f5d9077c9069c21e5ffb88e8f2936638243d2a"} Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.165903 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:13 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:13 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:13 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.165947 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:13 crc kubenswrapper[4806]: I1127 10:24:13.255658 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" podStartSLOduration=136.255642346 podStartE2EDuration="2m16.255642346s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:13.203336663 +0000 UTC m=+157.789927427" watchObservedRunningTime="2025-11-27 10:24:13.255642346 +0000 UTC m=+157.842233110" Nov 27 10:24:14 crc kubenswrapper[4806]: I1127 10:24:14.163590 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:14 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:14 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:14 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:14 crc kubenswrapper[4806]: I1127 10:24:14.164010 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:14 crc kubenswrapper[4806]: I1127 10:24:14.195743 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3b373728-67ae-4cf8-bd5a-0678e7123d46","Type":"ContainerStarted","Data":"2ca49a35db0dfe2c75e54d039179fdb5b0ce49d38fa726b8f9cf12eeff6f60ca"} Nov 27 10:24:14 crc kubenswrapper[4806]: I1127 10:24:14.214700 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.214681563 podStartE2EDuration="3.214681563s" podCreationTimestamp="2025-11-27 10:24:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:14.213814409 +0000 UTC m=+158.800405173" watchObservedRunningTime="2025-11-27 10:24:14.214681563 +0000 UTC m=+158.801272327" Nov 27 10:24:14 crc kubenswrapper[4806]: I1127 10:24:14.450426 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:24:14 crc kubenswrapper[4806]: I1127 10:24:14.450534 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:24:14 crc kubenswrapper[4806]: I1127 10:24:14.634105 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 10:24:14 crc kubenswrapper[4806]: I1127 10:24:14.776123 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66dfffe4-4026-409f-a8a2-bc8d23c55ba1-kubelet-dir\") pod \"66dfffe4-4026-409f-a8a2-bc8d23c55ba1\" (UID: \"66dfffe4-4026-409f-a8a2-bc8d23c55ba1\") " Nov 27 10:24:14 crc kubenswrapper[4806]: I1127 10:24:14.776815 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66dfffe4-4026-409f-a8a2-bc8d23c55ba1-kube-api-access\") pod \"66dfffe4-4026-409f-a8a2-bc8d23c55ba1\" (UID: \"66dfffe4-4026-409f-a8a2-bc8d23c55ba1\") " Nov 27 10:24:14 crc kubenswrapper[4806]: I1127 10:24:14.776216 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66dfffe4-4026-409f-a8a2-bc8d23c55ba1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "66dfffe4-4026-409f-a8a2-bc8d23c55ba1" (UID: "66dfffe4-4026-409f-a8a2-bc8d23c55ba1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:24:14 crc kubenswrapper[4806]: I1127 10:24:14.777528 4806 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66dfffe4-4026-409f-a8a2-bc8d23c55ba1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 10:24:14 crc kubenswrapper[4806]: I1127 10:24:14.790756 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66dfffe4-4026-409f-a8a2-bc8d23c55ba1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "66dfffe4-4026-409f-a8a2-bc8d23c55ba1" (UID: "66dfffe4-4026-409f-a8a2-bc8d23c55ba1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:24:14 crc kubenswrapper[4806]: I1127 10:24:14.887015 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66dfffe4-4026-409f-a8a2-bc8d23c55ba1-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 10:24:15 crc kubenswrapper[4806]: I1127 10:24:15.164860 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:15 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:15 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:15 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:15 crc kubenswrapper[4806]: I1127 10:24:15.164991 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:15 crc kubenswrapper[4806]: I1127 10:24:15.223276 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 27 10:24:15 crc kubenswrapper[4806]: I1127 10:24:15.223262 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"66dfffe4-4026-409f-a8a2-bc8d23c55ba1","Type":"ContainerDied","Data":"fd68911ac3e6e9f4777c617d892b72e20c844327de1908f811db350e87fa80fc"} Nov 27 10:24:15 crc kubenswrapper[4806]: I1127 10:24:15.223367 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd68911ac3e6e9f4777c617d892b72e20c844327de1908f811db350e87fa80fc" Nov 27 10:24:15 crc kubenswrapper[4806]: I1127 10:24:15.235251 4806 generic.go:334] "Generic (PLEG): container finished" podID="3b373728-67ae-4cf8-bd5a-0678e7123d46" containerID="2ca49a35db0dfe2c75e54d039179fdb5b0ce49d38fa726b8f9cf12eeff6f60ca" exitCode=0 Nov 27 10:24:15 crc kubenswrapper[4806]: I1127 10:24:15.235316 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3b373728-67ae-4cf8-bd5a-0678e7123d46","Type":"ContainerDied","Data":"2ca49a35db0dfe2c75e54d039179fdb5b0ce49d38fa726b8f9cf12eeff6f60ca"} Nov 27 10:24:16 crc kubenswrapper[4806]: I1127 10:24:16.186701 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:16 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:16 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:16 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:16 crc kubenswrapper[4806]: I1127 10:24:16.186784 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:16 crc kubenswrapper[4806]: I1127 10:24:16.794881 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 10:24:16 crc kubenswrapper[4806]: I1127 10:24:16.947665 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b373728-67ae-4cf8-bd5a-0678e7123d46-kubelet-dir\") pod \"3b373728-67ae-4cf8-bd5a-0678e7123d46\" (UID: \"3b373728-67ae-4cf8-bd5a-0678e7123d46\") " Nov 27 10:24:16 crc kubenswrapper[4806]: I1127 10:24:16.947764 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b373728-67ae-4cf8-bd5a-0678e7123d46-kube-api-access\") pod \"3b373728-67ae-4cf8-bd5a-0678e7123d46\" (UID: \"3b373728-67ae-4cf8-bd5a-0678e7123d46\") " Nov 27 10:24:16 crc kubenswrapper[4806]: I1127 10:24:16.948855 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b373728-67ae-4cf8-bd5a-0678e7123d46-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3b373728-67ae-4cf8-bd5a-0678e7123d46" (UID: "3b373728-67ae-4cf8-bd5a-0678e7123d46"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:24:16 crc kubenswrapper[4806]: I1127 10:24:16.979039 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b373728-67ae-4cf8-bd5a-0678e7123d46-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3b373728-67ae-4cf8-bd5a-0678e7123d46" (UID: "3b373728-67ae-4cf8-bd5a-0678e7123d46"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.049169 4806 patch_prober.go:28] interesting pod/downloads-7954f5f757-9ppfs container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.049278 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9ppfs" podUID="52dd6d6d-3a15-4467-9d82-37fd2fa5794b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.050432 4806 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3b373728-67ae-4cf8-bd5a-0678e7123d46-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.050508 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b373728-67ae-4cf8-bd5a-0678e7123d46-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.050572 4806 patch_prober.go:28] interesting pod/downloads-7954f5f757-9ppfs container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.050728 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9ppfs" podUID="52dd6d6d-3a15-4467-9d82-37fd2fa5794b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.165783 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:17 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:17 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:17 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.165845 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.335897 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3b373728-67ae-4cf8-bd5a-0678e7123d46","Type":"ContainerDied","Data":"1f6bc37a304a1f08ed124f8ffc6a45924624dbecad6144ebe6a7dfb4bcf7b72f"} Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.335943 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f6bc37a304a1f08ed124f8ffc6a45924624dbecad6144ebe6a7dfb4bcf7b72f" Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.335977 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.377259 4806 generic.go:334] "Generic (PLEG): container finished" podID="7308dfe6-f141-4f20-a966-075d17882333" containerID="535ec0a3c11c7d698bab89bdd18508c2a1ab57d28df16f3cf0b6cd72803ce638" exitCode=0 Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.377562 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" event={"ID":"7308dfe6-f141-4f20-a966-075d17882333","Type":"ContainerDied","Data":"535ec0a3c11c7d698bab89bdd18508c2a1ab57d28df16f3cf0b6cd72803ce638"} Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.411407 4806 patch_prober.go:28] interesting pod/console-f9d7485db-6fqzk container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 27 10:24:17 crc kubenswrapper[4806]: I1127 10:24:17.411474 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6fqzk" podUID="97042267-61ca-4864-a50a-60587abc7619" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 27 10:24:17 crc kubenswrapper[4806]: E1127 10:24:17.474579 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-pod3b373728_67ae_4cf8_bd5a_0678e7123d46.slice/crio-1f6bc37a304a1f08ed124f8ffc6a45924624dbecad6144ebe6a7dfb4bcf7b72f\": RecentStats: unable to find data in memory cache]" Nov 27 10:24:18 crc kubenswrapper[4806]: I1127 10:24:18.211701 4806 patch_prober.go:28] interesting pod/router-default-5444994796-kpjrm container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 27 10:24:18 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Nov 27 10:24:18 crc kubenswrapper[4806]: [+]process-running ok Nov 27 10:24:18 crc kubenswrapper[4806]: healthz check failed Nov 27 10:24:18 crc kubenswrapper[4806]: I1127 10:24:18.211793 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-kpjrm" podUID="cc69668d-b5d6-452a-99bf-b9a14c90a65d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 27 10:24:18 crc kubenswrapper[4806]: I1127 10:24:18.875275 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.035614 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7308dfe6-f141-4f20-a966-075d17882333-secret-volume\") pod \"7308dfe6-f141-4f20-a966-075d17882333\" (UID: \"7308dfe6-f141-4f20-a966-075d17882333\") " Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.035715 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbs8q\" (UniqueName: \"kubernetes.io/projected/7308dfe6-f141-4f20-a966-075d17882333-kube-api-access-rbs8q\") pod \"7308dfe6-f141-4f20-a966-075d17882333\" (UID: \"7308dfe6-f141-4f20-a966-075d17882333\") " Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.035797 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7308dfe6-f141-4f20-a966-075d17882333-config-volume\") pod \"7308dfe6-f141-4f20-a966-075d17882333\" (UID: \"7308dfe6-f141-4f20-a966-075d17882333\") " Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.038599 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7308dfe6-f141-4f20-a966-075d17882333-config-volume" (OuterVolumeSpecName: "config-volume") pod "7308dfe6-f141-4f20-a966-075d17882333" (UID: "7308dfe6-f141-4f20-a966-075d17882333"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.047157 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7308dfe6-f141-4f20-a966-075d17882333-kube-api-access-rbs8q" (OuterVolumeSpecName: "kube-api-access-rbs8q") pod "7308dfe6-f141-4f20-a966-075d17882333" (UID: "7308dfe6-f141-4f20-a966-075d17882333"). InnerVolumeSpecName "kube-api-access-rbs8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.076203 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7308dfe6-f141-4f20-a966-075d17882333-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7308dfe6-f141-4f20-a966-075d17882333" (UID: "7308dfe6-f141-4f20-a966-075d17882333"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.151916 4806 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7308dfe6-f141-4f20-a966-075d17882333-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.151958 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbs8q\" (UniqueName: \"kubernetes.io/projected/7308dfe6-f141-4f20-a966-075d17882333-kube-api-access-rbs8q\") on node \"crc\" DevicePath \"\"" Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.151971 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7308dfe6-f141-4f20-a966-075d17882333-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.176981 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.189433 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-kpjrm" Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.509738 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.510831 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk" event={"ID":"7308dfe6-f141-4f20-a966-075d17882333","Type":"ContainerDied","Data":"3d89ba4efb288e12349ee7d6e13a7001021ceca7eae9fb2e851e093a9a19da3e"} Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.510916 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d89ba4efb288e12349ee7d6e13a7001021ceca7eae9fb2e851e093a9a19da3e" Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.875165 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:24:19 crc kubenswrapper[4806]: I1127 10:24:19.884849 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e-metrics-certs\") pod \"network-metrics-daemon-p2gpw\" (UID: \"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e\") " pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:24:20 crc kubenswrapper[4806]: I1127 10:24:20.140277 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p2gpw" Nov 27 10:24:21 crc kubenswrapper[4806]: I1127 10:24:21.004145 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-p2gpw"] Nov 27 10:24:21 crc kubenswrapper[4806]: W1127 10:24:21.089574 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fbf13a5_7d98_47f6_a2b7_eee631e5eb4e.slice/crio-0d13f3c26dc9ea992c486b7c07aab8244ab04421aa916ca4d4a1f25a03e83e8b WatchSource:0}: Error finding container 0d13f3c26dc9ea992c486b7c07aab8244ab04421aa916ca4d4a1f25a03e83e8b: Status 404 returned error can't find the container with id 0d13f3c26dc9ea992c486b7c07aab8244ab04421aa916ca4d4a1f25a03e83e8b Nov 27 10:24:21 crc kubenswrapper[4806]: I1127 10:24:21.618703 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" event={"ID":"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e","Type":"ContainerStarted","Data":"0d13f3c26dc9ea992c486b7c07aab8244ab04421aa916ca4d4a1f25a03e83e8b"} Nov 27 10:24:22 crc kubenswrapper[4806]: I1127 10:24:22.632739 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" event={"ID":"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e","Type":"ContainerStarted","Data":"907edf3c1976b06f4a19a3f01b1004231f0f3ee90dd772587ec577afe0feb4c1"} Nov 27 10:24:27 crc kubenswrapper[4806]: I1127 10:24:27.053995 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-9ppfs" Nov 27 10:24:27 crc kubenswrapper[4806]: I1127 10:24:27.415790 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:24:27 crc kubenswrapper[4806]: I1127 10:24:27.429901 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:24:30 crc kubenswrapper[4806]: I1127 10:24:30.698033 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:24:36 crc kubenswrapper[4806]: I1127 10:24:36.342186 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p2gpw" event={"ID":"3fbf13a5-7d98-47f6-a2b7-eee631e5eb4e","Type":"ContainerStarted","Data":"eaae3e4655e15a621295fa851dd7009ae9065d427d648b39bfbeb644bf29a913"} Nov 27 10:24:37 crc kubenswrapper[4806]: I1127 10:24:37.879927 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sg6x8" Nov 27 10:24:37 crc kubenswrapper[4806]: I1127 10:24:37.915028 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-p2gpw" podStartSLOduration=160.914489646 podStartE2EDuration="2m40.914489646s" podCreationTimestamp="2025-11-27 10:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:24:37.366287791 +0000 UTC m=+181.952878565" watchObservedRunningTime="2025-11-27 10:24:37.914489646 +0000 UTC m=+182.501080410" Nov 27 10:24:44 crc kubenswrapper[4806]: I1127 10:24:44.047630 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 27 10:24:44 crc kubenswrapper[4806]: I1127 10:24:44.450626 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:24:44 crc kubenswrapper[4806]: I1127 10:24:44.450968 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:24:46 crc kubenswrapper[4806]: I1127 10:24:46.610111 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l956q"] Nov 27 10:24:50 crc kubenswrapper[4806]: E1127 10:24:50.362573 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 27 10:24:50 crc kubenswrapper[4806]: E1127 10:24:50.364172 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8z57v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-l4x5q_openshift-marketplace(8cb0392a-70a2-48c0-a69e-4421dbefe86e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 10:24:50 crc kubenswrapper[4806]: E1127 10:24:50.365645 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-l4x5q" podUID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" Nov 27 10:24:50 crc kubenswrapper[4806]: E1127 10:24:50.452943 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 27 10:24:50 crc kubenswrapper[4806]: E1127 10:24:50.453192 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7mz5t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-2cthc_openshift-marketplace(ae461a45-1190-46fc-acf5-8954b951b25e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 10:24:50 crc kubenswrapper[4806]: E1127 10:24:50.456953 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-2cthc" podUID="ae461a45-1190-46fc-acf5-8954b951b25e" Nov 27 10:24:50 crc kubenswrapper[4806]: E1127 10:24:50.472259 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 27 10:24:50 crc kubenswrapper[4806]: E1127 10:24:50.472505 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tf426,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-wfw97_openshift-marketplace(ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 10:24:50 crc kubenswrapper[4806]: E1127 10:24:50.474109 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-wfw97" podUID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" Nov 27 10:24:50 crc kubenswrapper[4806]: E1127 10:24:50.498772 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-l4x5q" podUID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" Nov 27 10:24:50 crc kubenswrapper[4806]: E1127 10:24:50.535013 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 27 10:24:50 crc kubenswrapper[4806]: E1127 10:24:50.535358 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w5qgh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-gp6rt_openshift-marketplace(3209cb7e-221c-4c6a-9864-8101e8171078): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 10:24:50 crc kubenswrapper[4806]: E1127 10:24:50.537608 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-gp6rt" podUID="3209cb7e-221c-4c6a-9864-8101e8171078" Nov 27 10:24:51 crc kubenswrapper[4806]: I1127 10:24:51.464587 4806 generic.go:334] "Generic (PLEG): container finished" podID="a3bc679e-e16d-47ce-9015-c110b1604ac1" containerID="2120d4c87d3cd8a3b7f1b50c681960d0f4f20ee99282202ad4bff85e20c411db" exitCode=0 Nov 27 10:24:51 crc kubenswrapper[4806]: I1127 10:24:51.464651 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jprw7" event={"ID":"a3bc679e-e16d-47ce-9015-c110b1604ac1","Type":"ContainerDied","Data":"2120d4c87d3cd8a3b7f1b50c681960d0f4f20ee99282202ad4bff85e20c411db"} Nov 27 10:24:51 crc kubenswrapper[4806]: I1127 10:24:51.471974 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nmrrm" event={"ID":"293a1c8c-8131-462c-9dc4-4a8c516478c5","Type":"ContainerStarted","Data":"db6a9629a43f497636805a0771797b16ca4fca530a30ce38976b5bff602fe326"} Nov 27 10:24:51 crc kubenswrapper[4806]: I1127 10:24:51.476590 4806 generic.go:334] "Generic (PLEG): container finished" podID="72810aad-3a98-4ed6-bf65-6434e72770fb" containerID="ba5872983a56578cc4e17349dedf2fe3b841dbdc0d9bb307e89d4523a23c6a8d" exitCode=0 Nov 27 10:24:51 crc kubenswrapper[4806]: I1127 10:24:51.476718 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxtlg" event={"ID":"72810aad-3a98-4ed6-bf65-6434e72770fb","Type":"ContainerDied","Data":"ba5872983a56578cc4e17349dedf2fe3b841dbdc0d9bb307e89d4523a23c6a8d"} Nov 27 10:24:51 crc kubenswrapper[4806]: I1127 10:24:51.494543 4806 generic.go:334] "Generic (PLEG): container finished" podID="c3c3761a-7b4a-4053-bf59-0634f90d1e91" containerID="8909c19cb3af9cf529da6b2adca42a99919ebf6b1938f3081abdbd49fa9f6af4" exitCode=0 Nov 27 10:24:51 crc kubenswrapper[4806]: I1127 10:24:51.496496 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9v4rb" event={"ID":"c3c3761a-7b4a-4053-bf59-0634f90d1e91","Type":"ContainerDied","Data":"8909c19cb3af9cf529da6b2adca42a99919ebf6b1938f3081abdbd49fa9f6af4"} Nov 27 10:24:51 crc kubenswrapper[4806]: E1127 10:24:51.497630 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2cthc" podUID="ae461a45-1190-46fc-acf5-8954b951b25e" Nov 27 10:24:51 crc kubenswrapper[4806]: E1127 10:24:51.497869 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-wfw97" podUID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" Nov 27 10:24:51 crc kubenswrapper[4806]: E1127 10:24:51.523781 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-gp6rt" podUID="3209cb7e-221c-4c6a-9864-8101e8171078" Nov 27 10:24:52 crc kubenswrapper[4806]: I1127 10:24:52.536216 4806 generic.go:334] "Generic (PLEG): container finished" podID="293a1c8c-8131-462c-9dc4-4a8c516478c5" containerID="db6a9629a43f497636805a0771797b16ca4fca530a30ce38976b5bff602fe326" exitCode=0 Nov 27 10:24:52 crc kubenswrapper[4806]: I1127 10:24:52.536291 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nmrrm" event={"ID":"293a1c8c-8131-462c-9dc4-4a8c516478c5","Type":"ContainerDied","Data":"db6a9629a43f497636805a0771797b16ca4fca530a30ce38976b5bff602fe326"} Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.557249 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nmrrm" event={"ID":"293a1c8c-8131-462c-9dc4-4a8c516478c5","Type":"ContainerStarted","Data":"fd8d33d4b580fcaab375582ed651c86fc547e475440aca5e7c5716bcbdd71bcb"} Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.559389 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxtlg" event={"ID":"72810aad-3a98-4ed6-bf65-6434e72770fb","Type":"ContainerStarted","Data":"87d26c6660e77d8d09587c122f04bb3a4f17cde8c6361579182bd8a012ce4236"} Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.561001 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9v4rb" event={"ID":"c3c3761a-7b4a-4053-bf59-0634f90d1e91","Type":"ContainerStarted","Data":"d6d293d72fab515bfcef36b14063f6a74895a0e88874855d0062a5a2d976c33f"} Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.567092 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jprw7" event={"ID":"a3bc679e-e16d-47ce-9015-c110b1604ac1","Type":"ContainerStarted","Data":"f5ff479bfdd746688aa212d0a72af52cbc5e3451351b9e28b812d6229ade8015"} Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.587849 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nmrrm" podStartSLOduration=4.181994142 podStartE2EDuration="45.587825852s" podCreationTimestamp="2025-11-27 10:24:10 +0000 UTC" firstStartedPulling="2025-11-27 10:24:13.166473462 +0000 UTC m=+157.753064226" lastFinishedPulling="2025-11-27 10:24:54.572305172 +0000 UTC m=+199.158895936" observedRunningTime="2025-11-27 10:24:55.583928801 +0000 UTC m=+200.170519575" watchObservedRunningTime="2025-11-27 10:24:55.587825852 +0000 UTC m=+200.174416616" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.609167 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9v4rb" podStartSLOduration=6.097083176 podStartE2EDuration="48.609139273s" podCreationTimestamp="2025-11-27 10:24:07 +0000 UTC" firstStartedPulling="2025-11-27 10:24:11.959763686 +0000 UTC m=+156.546354450" lastFinishedPulling="2025-11-27 10:24:54.471819783 +0000 UTC m=+199.058410547" observedRunningTime="2025-11-27 10:24:55.607226468 +0000 UTC m=+200.193817232" watchObservedRunningTime="2025-11-27 10:24:55.609139273 +0000 UTC m=+200.195730037" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.627898 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jprw7" podStartSLOduration=4.803835165 podStartE2EDuration="49.627870049s" podCreationTimestamp="2025-11-27 10:24:06 +0000 UTC" firstStartedPulling="2025-11-27 10:24:09.591417025 +0000 UTC m=+154.178007789" lastFinishedPulling="2025-11-27 10:24:54.415451909 +0000 UTC m=+199.002042673" observedRunningTime="2025-11-27 10:24:55.627177329 +0000 UTC m=+200.213768103" watchObservedRunningTime="2025-11-27 10:24:55.627870049 +0000 UTC m=+200.214460813" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.654785 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nxtlg" podStartSLOduration=5.394545871 podStartE2EDuration="46.65476188s" podCreationTimestamp="2025-11-27 10:24:09 +0000 UTC" firstStartedPulling="2025-11-27 10:24:13.166173763 +0000 UTC m=+157.752764527" lastFinishedPulling="2025-11-27 10:24:54.426389772 +0000 UTC m=+199.012980536" observedRunningTime="2025-11-27 10:24:55.651911858 +0000 UTC m=+200.238502622" watchObservedRunningTime="2025-11-27 10:24:55.65476188 +0000 UTC m=+200.241352644" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.802925 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 27 10:24:55 crc kubenswrapper[4806]: E1127 10:24:55.803177 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b373728-67ae-4cf8-bd5a-0678e7123d46" containerName="pruner" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.803191 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b373728-67ae-4cf8-bd5a-0678e7123d46" containerName="pruner" Nov 27 10:24:55 crc kubenswrapper[4806]: E1127 10:24:55.803207 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66dfffe4-4026-409f-a8a2-bc8d23c55ba1" containerName="pruner" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.803213 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="66dfffe4-4026-409f-a8a2-bc8d23c55ba1" containerName="pruner" Nov 27 10:24:55 crc kubenswrapper[4806]: E1127 10:24:55.803227 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7308dfe6-f141-4f20-a966-075d17882333" containerName="collect-profiles" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.803262 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7308dfe6-f141-4f20-a966-075d17882333" containerName="collect-profiles" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.803375 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="66dfffe4-4026-409f-a8a2-bc8d23c55ba1" containerName="pruner" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.803386 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="7308dfe6-f141-4f20-a966-075d17882333" containerName="collect-profiles" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.803398 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b373728-67ae-4cf8-bd5a-0678e7123d46" containerName="pruner" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.804342 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.807207 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.807459 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.823304 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.991520 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/10db5665-dc97-4ee6-88b3-86ca80aae85c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"10db5665-dc97-4ee6-88b3-86ca80aae85c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 10:24:55 crc kubenswrapper[4806]: I1127 10:24:55.992069 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10db5665-dc97-4ee6-88b3-86ca80aae85c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"10db5665-dc97-4ee6-88b3-86ca80aae85c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 10:24:56 crc kubenswrapper[4806]: I1127 10:24:56.093345 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/10db5665-dc97-4ee6-88b3-86ca80aae85c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"10db5665-dc97-4ee6-88b3-86ca80aae85c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 10:24:56 crc kubenswrapper[4806]: I1127 10:24:56.093440 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10db5665-dc97-4ee6-88b3-86ca80aae85c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"10db5665-dc97-4ee6-88b3-86ca80aae85c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 10:24:56 crc kubenswrapper[4806]: I1127 10:24:56.093847 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/10db5665-dc97-4ee6-88b3-86ca80aae85c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"10db5665-dc97-4ee6-88b3-86ca80aae85c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 10:24:56 crc kubenswrapper[4806]: I1127 10:24:56.131312 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10db5665-dc97-4ee6-88b3-86ca80aae85c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"10db5665-dc97-4ee6-88b3-86ca80aae85c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 10:24:56 crc kubenswrapper[4806]: I1127 10:24:56.423485 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 10:24:56 crc kubenswrapper[4806]: I1127 10:24:56.963462 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 27 10:24:56 crc kubenswrapper[4806]: W1127 10:24:56.979881 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod10db5665_dc97_4ee6_88b3_86ca80aae85c.slice/crio-38d8c74421774e28b592b6becf8d14026b3ce02c04076373bbd621c0821e5521 WatchSource:0}: Error finding container 38d8c74421774e28b592b6becf8d14026b3ce02c04076373bbd621c0821e5521: Status 404 returned error can't find the container with id 38d8c74421774e28b592b6becf8d14026b3ce02c04076373bbd621c0821e5521 Nov 27 10:24:57 crc kubenswrapper[4806]: I1127 10:24:57.030466 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:24:57 crc kubenswrapper[4806]: I1127 10:24:57.030532 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:24:57 crc kubenswrapper[4806]: I1127 10:24:57.583112 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"10db5665-dc97-4ee6-88b3-86ca80aae85c","Type":"ContainerStarted","Data":"38d8c74421774e28b592b6becf8d14026b3ce02c04076373bbd621c0821e5521"} Nov 27 10:24:57 crc kubenswrapper[4806]: I1127 10:24:57.776889 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:24:57 crc kubenswrapper[4806]: I1127 10:24:57.776940 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:24:58 crc kubenswrapper[4806]: I1127 10:24:58.483156 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jprw7" podUID="a3bc679e-e16d-47ce-9015-c110b1604ac1" containerName="registry-server" probeResult="failure" output=< Nov 27 10:24:58 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 10:24:58 crc kubenswrapper[4806]: > Nov 27 10:24:58 crc kubenswrapper[4806]: I1127 10:24:58.589536 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"10db5665-dc97-4ee6-88b3-86ca80aae85c","Type":"ContainerStarted","Data":"51ffa67e6024006074ed69cc3d98a8df79f3735c192db4cc3fa9d8781c98d26b"} Nov 27 10:24:58 crc kubenswrapper[4806]: I1127 10:24:58.821998 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-9v4rb" podUID="c3c3761a-7b4a-4053-bf59-0634f90d1e91" containerName="registry-server" probeResult="failure" output=< Nov 27 10:24:58 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 10:24:58 crc kubenswrapper[4806]: > Nov 27 10:24:59 crc kubenswrapper[4806]: I1127 10:24:59.596653 4806 generic.go:334] "Generic (PLEG): container finished" podID="10db5665-dc97-4ee6-88b3-86ca80aae85c" containerID="51ffa67e6024006074ed69cc3d98a8df79f3735c192db4cc3fa9d8781c98d26b" exitCode=0 Nov 27 10:24:59 crc kubenswrapper[4806]: I1127 10:24:59.596934 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"10db5665-dc97-4ee6-88b3-86ca80aae85c","Type":"ContainerDied","Data":"51ffa67e6024006074ed69cc3d98a8df79f3735c192db4cc3fa9d8781c98d26b"} Nov 27 10:25:00 crc kubenswrapper[4806]: I1127 10:25:00.138115 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:25:00 crc kubenswrapper[4806]: I1127 10:25:00.138184 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:25:00 crc kubenswrapper[4806]: I1127 10:25:00.268279 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:25:00 crc kubenswrapper[4806]: I1127 10:25:00.654283 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:25:00 crc kubenswrapper[4806]: I1127 10:25:00.776131 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:25:00 crc kubenswrapper[4806]: I1127 10:25:00.777005 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:25:00 crc kubenswrapper[4806]: I1127 10:25:00.925500 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.070916 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10db5665-dc97-4ee6-88b3-86ca80aae85c-kube-api-access\") pod \"10db5665-dc97-4ee6-88b3-86ca80aae85c\" (UID: \"10db5665-dc97-4ee6-88b3-86ca80aae85c\") " Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.071000 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/10db5665-dc97-4ee6-88b3-86ca80aae85c-kubelet-dir\") pod \"10db5665-dc97-4ee6-88b3-86ca80aae85c\" (UID: \"10db5665-dc97-4ee6-88b3-86ca80aae85c\") " Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.071144 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10db5665-dc97-4ee6-88b3-86ca80aae85c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "10db5665-dc97-4ee6-88b3-86ca80aae85c" (UID: "10db5665-dc97-4ee6-88b3-86ca80aae85c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.071416 4806 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/10db5665-dc97-4ee6-88b3-86ca80aae85c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.090286 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10db5665-dc97-4ee6-88b3-86ca80aae85c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "10db5665-dc97-4ee6-88b3-86ca80aae85c" (UID: "10db5665-dc97-4ee6-88b3-86ca80aae85c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.172704 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10db5665-dc97-4ee6-88b3-86ca80aae85c-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.337279 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2szrd"] Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.337572 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" podUID="dfd928ce-640d-4b02-9d41-a81a12b3fbfc" containerName="controller-manager" containerID="cri-o://88784f783a81f6f615b47abf4bab6fc4290e752297efcf5b66c9ea897f0c621b" gracePeriod=30 Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.474073 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk"] Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.474366 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" podUID="9369b0d5-ab28-4cc3-a500-c8efebdbdc1f" containerName="route-controller-manager" containerID="cri-o://ea22f82fa7fcdf63523429c14e16ee4ed5a354fc2536a317d148606319f95d4c" gracePeriod=30 Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.612517 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.612681 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"10db5665-dc97-4ee6-88b3-86ca80aae85c","Type":"ContainerDied","Data":"38d8c74421774e28b592b6becf8d14026b3ce02c04076373bbd621c0821e5521"} Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.613639 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38d8c74421774e28b592b6becf8d14026b3ce02c04076373bbd621c0821e5521" Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.704071 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxtlg"] Nov 27 10:25:01 crc kubenswrapper[4806]: I1127 10:25:01.859436 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nmrrm" podUID="293a1c8c-8131-462c-9dc4-4a8c516478c5" containerName="registry-server" probeResult="failure" output=< Nov 27 10:25:01 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 10:25:01 crc kubenswrapper[4806]: > Nov 27 10:25:02 crc kubenswrapper[4806]: I1127 10:25:02.636316 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" event={"ID":"dfd928ce-640d-4b02-9d41-a81a12b3fbfc","Type":"ContainerDied","Data":"88784f783a81f6f615b47abf4bab6fc4290e752297efcf5b66c9ea897f0c621b"} Nov 27 10:25:02 crc kubenswrapper[4806]: I1127 10:25:02.636414 4806 generic.go:334] "Generic (PLEG): container finished" podID="dfd928ce-640d-4b02-9d41-a81a12b3fbfc" containerID="88784f783a81f6f615b47abf4bab6fc4290e752297efcf5b66c9ea897f0c621b" exitCode=0 Nov 27 10:25:02 crc kubenswrapper[4806]: I1127 10:25:02.640267 4806 generic.go:334] "Generic (PLEG): container finished" podID="9369b0d5-ab28-4cc3-a500-c8efebdbdc1f" containerID="ea22f82fa7fcdf63523429c14e16ee4ed5a354fc2536a317d148606319f95d4c" exitCode=0 Nov 27 10:25:02 crc kubenswrapper[4806]: I1127 10:25:02.641103 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nxtlg" podUID="72810aad-3a98-4ed6-bf65-6434e72770fb" containerName="registry-server" containerID="cri-o://87d26c6660e77d8d09587c122f04bb3a4f17cde8c6361579182bd8a012ce4236" gracePeriod=2 Nov 27 10:25:02 crc kubenswrapper[4806]: I1127 10:25:02.644523 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" event={"ID":"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f","Type":"ContainerDied","Data":"ea22f82fa7fcdf63523429c14e16ee4ed5a354fc2536a317d148606319f95d4c"} Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.431103 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.475093 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-55568dc997-pmlhd"] Nov 27 10:25:03 crc kubenswrapper[4806]: E1127 10:25:03.475364 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfd928ce-640d-4b02-9d41-a81a12b3fbfc" containerName="controller-manager" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.475380 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfd928ce-640d-4b02-9d41-a81a12b3fbfc" containerName="controller-manager" Nov 27 10:25:03 crc kubenswrapper[4806]: E1127 10:25:03.475403 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10db5665-dc97-4ee6-88b3-86ca80aae85c" containerName="pruner" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.475409 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="10db5665-dc97-4ee6-88b3-86ca80aae85c" containerName="pruner" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.475506 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfd928ce-640d-4b02-9d41-a81a12b3fbfc" containerName="controller-manager" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.475519 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="10db5665-dc97-4ee6-88b3-86ca80aae85c" containerName="pruner" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.475956 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.499070 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55568dc997-pmlhd"] Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.598829 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.616986 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-serving-cert\") pod \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.617086 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqnvq\" (UniqueName: \"kubernetes.io/projected/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-kube-api-access-hqnvq\") pod \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.617197 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-config\") pod \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.617252 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-proxy-ca-bundles\") pod \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.617273 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-client-ca\") pod \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\" (UID: \"dfd928ce-640d-4b02-9d41-a81a12b3fbfc\") " Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.617518 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-proxy-ca-bundles\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.617554 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d09a5847-2755-42b2-a220-a3b254b3ba19-serving-cert\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.617602 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n848\" (UniqueName: \"kubernetes.io/projected/d09a5847-2755-42b2-a220-a3b254b3ba19-kube-api-access-2n848\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.617644 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-config\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.617667 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-client-ca\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.619677 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-client-ca" (OuterVolumeSpecName: "client-ca") pod "dfd928ce-640d-4b02-9d41-a81a12b3fbfc" (UID: "dfd928ce-640d-4b02-9d41-a81a12b3fbfc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.620357 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-config" (OuterVolumeSpecName: "config") pod "dfd928ce-640d-4b02-9d41-a81a12b3fbfc" (UID: "dfd928ce-640d-4b02-9d41-a81a12b3fbfc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.620657 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "dfd928ce-640d-4b02-9d41-a81a12b3fbfc" (UID: "dfd928ce-640d-4b02-9d41-a81a12b3fbfc"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.637511 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dfd928ce-640d-4b02-9d41-a81a12b3fbfc" (UID: "dfd928ce-640d-4b02-9d41-a81a12b3fbfc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.639070 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-kube-api-access-hqnvq" (OuterVolumeSpecName: "kube-api-access-hqnvq") pod "dfd928ce-640d-4b02-9d41-a81a12b3fbfc" (UID: "dfd928ce-640d-4b02-9d41-a81a12b3fbfc"). InnerVolumeSpecName "kube-api-access-hqnvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.658815 4806 generic.go:334] "Generic (PLEG): container finished" podID="72810aad-3a98-4ed6-bf65-6434e72770fb" containerID="87d26c6660e77d8d09587c122f04bb3a4f17cde8c6361579182bd8a012ce4236" exitCode=0 Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.658922 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxtlg" event={"ID":"72810aad-3a98-4ed6-bf65-6434e72770fb","Type":"ContainerDied","Data":"87d26c6660e77d8d09587c122f04bb3a4f17cde8c6361579182bd8a012ce4236"} Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.660949 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" event={"ID":"dfd928ce-640d-4b02-9d41-a81a12b3fbfc","Type":"ContainerDied","Data":"89a15517e56ec66939df665561ceecac36e45c434e969c23ffa2dd070ba4855c"} Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.660986 4806 scope.go:117] "RemoveContainer" containerID="88784f783a81f6f615b47abf4bab6fc4290e752297efcf5b66c9ea897f0c621b" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.661190 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2szrd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.675384 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" event={"ID":"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f","Type":"ContainerDied","Data":"a67bcd61d986e56f9ded14e5b71955e79bef00bd9bf00ad29f44930ab6420ac4"} Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.675464 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.695933 4806 scope.go:117] "RemoveContainer" containerID="ea22f82fa7fcdf63523429c14e16ee4ed5a354fc2536a317d148606319f95d4c" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.720315 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-serving-cert\") pod \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.721514 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-config\") pod \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.721714 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krmjc\" (UniqueName: \"kubernetes.io/projected/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-kube-api-access-krmjc\") pod \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.721800 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-client-ca\") pod \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\" (UID: \"9369b0d5-ab28-4cc3-a500-c8efebdbdc1f\") " Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.722027 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-proxy-ca-bundles\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.722114 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d09a5847-2755-42b2-a220-a3b254b3ba19-serving-cert\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.721151 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2szrd"] Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.723334 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-client-ca" (OuterVolumeSpecName: "client-ca") pod "9369b0d5-ab28-4cc3-a500-c8efebdbdc1f" (UID: "9369b0d5-ab28-4cc3-a500-c8efebdbdc1f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.723375 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-config" (OuterVolumeSpecName: "config") pod "9369b0d5-ab28-4cc3-a500-c8efebdbdc1f" (UID: "9369b0d5-ab28-4cc3-a500-c8efebdbdc1f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.724955 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9369b0d5-ab28-4cc3-a500-c8efebdbdc1f" (UID: "9369b0d5-ab28-4cc3-a500-c8efebdbdc1f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.726146 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n848\" (UniqueName: \"kubernetes.io/projected/d09a5847-2755-42b2-a220-a3b254b3ba19-kube-api-access-2n848\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.726356 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-config\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.726409 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-client-ca\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.726479 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqnvq\" (UniqueName: \"kubernetes.io/projected/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-kube-api-access-hqnvq\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.726491 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.726518 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.726530 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.726539 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.726565 4806 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.726574 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.726584 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dfd928ce-640d-4b02-9d41-a81a12b3fbfc-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.726731 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-kube-api-access-krmjc" (OuterVolumeSpecName: "kube-api-access-krmjc") pod "9369b0d5-ab28-4cc3-a500-c8efebdbdc1f" (UID: "9369b0d5-ab28-4cc3-a500-c8efebdbdc1f"). InnerVolumeSpecName "kube-api-access-krmjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.727455 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-client-ca\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.728039 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-config\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.729819 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-proxy-ca-bundles\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.730578 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.732382 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2szrd"] Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.733598 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d09a5847-2755-42b2-a220-a3b254b3ba19-serving-cert\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.747952 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n848\" (UniqueName: \"kubernetes.io/projected/d09a5847-2755-42b2-a220-a3b254b3ba19-kube-api-access-2n848\") pod \"controller-manager-55568dc997-pmlhd\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.771110 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 27 10:25:03 crc kubenswrapper[4806]: E1127 10:25:03.771612 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72810aad-3a98-4ed6-bf65-6434e72770fb" containerName="registry-server" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.771683 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="72810aad-3a98-4ed6-bf65-6434e72770fb" containerName="registry-server" Nov 27 10:25:03 crc kubenswrapper[4806]: E1127 10:25:03.771740 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72810aad-3a98-4ed6-bf65-6434e72770fb" containerName="extract-content" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.771821 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="72810aad-3a98-4ed6-bf65-6434e72770fb" containerName="extract-content" Nov 27 10:25:03 crc kubenswrapper[4806]: E1127 10:25:03.771892 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9369b0d5-ab28-4cc3-a500-c8efebdbdc1f" containerName="route-controller-manager" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.771943 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="9369b0d5-ab28-4cc3-a500-c8efebdbdc1f" containerName="route-controller-manager" Nov 27 10:25:03 crc kubenswrapper[4806]: E1127 10:25:03.771994 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72810aad-3a98-4ed6-bf65-6434e72770fb" containerName="extract-utilities" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.772050 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="72810aad-3a98-4ed6-bf65-6434e72770fb" containerName="extract-utilities" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.772201 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="9369b0d5-ab28-4cc3-a500-c8efebdbdc1f" containerName="route-controller-manager" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.772286 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="72810aad-3a98-4ed6-bf65-6434e72770fb" containerName="registry-server" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.772755 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.792967 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.793750 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.799333 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.821375 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.827681 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krmjc\" (UniqueName: \"kubernetes.io/projected/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f-kube-api-access-krmjc\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.929075 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72810aad-3a98-4ed6-bf65-6434e72770fb-catalog-content\") pod \"72810aad-3a98-4ed6-bf65-6434e72770fb\" (UID: \"72810aad-3a98-4ed6-bf65-6434e72770fb\") " Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.929173 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72810aad-3a98-4ed6-bf65-6434e72770fb-utilities\") pod \"72810aad-3a98-4ed6-bf65-6434e72770fb\" (UID: \"72810aad-3a98-4ed6-bf65-6434e72770fb\") " Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.929276 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjqq7\" (UniqueName: \"kubernetes.io/projected/72810aad-3a98-4ed6-bf65-6434e72770fb-kube-api-access-mjqq7\") pod \"72810aad-3a98-4ed6-bf65-6434e72770fb\" (UID: \"72810aad-3a98-4ed6-bf65-6434e72770fb\") " Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.929724 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38a72bf5-a9ca-409a-b52d-b10ebefc316a-kube-api-access\") pod \"installer-9-crc\" (UID: \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.929758 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38a72bf5-a9ca-409a-b52d-b10ebefc316a-kubelet-dir\") pod \"installer-9-crc\" (UID: \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.929848 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/38a72bf5-a9ca-409a-b52d-b10ebefc316a-var-lock\") pod \"installer-9-crc\" (UID: \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.930760 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72810aad-3a98-4ed6-bf65-6434e72770fb-utilities" (OuterVolumeSpecName: "utilities") pod "72810aad-3a98-4ed6-bf65-6434e72770fb" (UID: "72810aad-3a98-4ed6-bf65-6434e72770fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.935416 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72810aad-3a98-4ed6-bf65-6434e72770fb-kube-api-access-mjqq7" (OuterVolumeSpecName: "kube-api-access-mjqq7") pod "72810aad-3a98-4ed6-bf65-6434e72770fb" (UID: "72810aad-3a98-4ed6-bf65-6434e72770fb"). InnerVolumeSpecName "kube-api-access-mjqq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:03 crc kubenswrapper[4806]: I1127 10:25:03.965584 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72810aad-3a98-4ed6-bf65-6434e72770fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72810aad-3a98-4ed6-bf65-6434e72770fb" (UID: "72810aad-3a98-4ed6-bf65-6434e72770fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.014906 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk"] Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.021726 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xb7gk"] Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.031729 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38a72bf5-a9ca-409a-b52d-b10ebefc316a-kube-api-access\") pod \"installer-9-crc\" (UID: \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.031793 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38a72bf5-a9ca-409a-b52d-b10ebefc316a-kubelet-dir\") pod \"installer-9-crc\" (UID: \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.031841 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/38a72bf5-a9ca-409a-b52d-b10ebefc316a-var-lock\") pod \"installer-9-crc\" (UID: \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.031901 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjqq7\" (UniqueName: \"kubernetes.io/projected/72810aad-3a98-4ed6-bf65-6434e72770fb-kube-api-access-mjqq7\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.031918 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72810aad-3a98-4ed6-bf65-6434e72770fb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.031930 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72810aad-3a98-4ed6-bf65-6434e72770fb-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.031988 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/38a72bf5-a9ca-409a-b52d-b10ebefc316a-var-lock\") pod \"installer-9-crc\" (UID: \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.033039 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38a72bf5-a9ca-409a-b52d-b10ebefc316a-kubelet-dir\") pod \"installer-9-crc\" (UID: \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.058916 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38a72bf5-a9ca-409a-b52d-b10ebefc316a-kube-api-access\") pod \"installer-9-crc\" (UID: \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.108837 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.132625 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9369b0d5-ab28-4cc3-a500-c8efebdbdc1f" path="/var/lib/kubelet/pods/9369b0d5-ab28-4cc3-a500-c8efebdbdc1f/volumes" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.133899 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfd928ce-640d-4b02-9d41-a81a12b3fbfc" path="/var/lib/kubelet/pods/dfd928ce-640d-4b02-9d41-a81a12b3fbfc/volumes" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.134737 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-55568dc997-pmlhd"] Nov 27 10:25:04 crc kubenswrapper[4806]: W1127 10:25:04.187795 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd09a5847_2755_42b2_a220_a3b254b3ba19.slice/crio-56c3f323e19e3b8cec4b6db5358e07079a88daad124f7b8cca5313b737618e88 WatchSource:0}: Error finding container 56c3f323e19e3b8cec4b6db5358e07079a88daad124f7b8cca5313b737618e88: Status 404 returned error can't find the container with id 56c3f323e19e3b8cec4b6db5358e07079a88daad124f7b8cca5313b737618e88 Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.497915 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 27 10:25:04 crc kubenswrapper[4806]: W1127 10:25:04.508136 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod38a72bf5_a9ca_409a_b52d_b10ebefc316a.slice/crio-44121b905ac36f2b6085d34a099ee85ccac28273e2187a74ce2a54dc8a689e25 WatchSource:0}: Error finding container 44121b905ac36f2b6085d34a099ee85ccac28273e2187a74ce2a54dc8a689e25: Status 404 returned error can't find the container with id 44121b905ac36f2b6085d34a099ee85ccac28273e2187a74ce2a54dc8a689e25 Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.688978 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxtlg" event={"ID":"72810aad-3a98-4ed6-bf65-6434e72770fb","Type":"ContainerDied","Data":"b79e6ab9132478dba2aec7a03522e189d5c0ecd284beb2c974db43a639bbdc5c"} Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.689054 4806 scope.go:117] "RemoveContainer" containerID="87d26c6660e77d8d09587c122f04bb3a4f17cde8c6361579182bd8a012ce4236" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.689256 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxtlg" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.700139 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4x5q" event={"ID":"8cb0392a-70a2-48c0-a69e-4421dbefe86e","Type":"ContainerStarted","Data":"6d847aa087b075d1492241d80b55736363187ba70efda8d87c94110a62a2c316"} Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.717666 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"38a72bf5-a9ca-409a-b52d-b10ebefc316a","Type":"ContainerStarted","Data":"44121b905ac36f2b6085d34a099ee85ccac28273e2187a74ce2a54dc8a689e25"} Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.721052 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" event={"ID":"d09a5847-2755-42b2-a220-a3b254b3ba19","Type":"ContainerStarted","Data":"9fd2be668fa8835cdd4446a034366d9183ad716747dab93d08d3aa690cd9ea15"} Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.721077 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" event={"ID":"d09a5847-2755-42b2-a220-a3b254b3ba19","Type":"ContainerStarted","Data":"56c3f323e19e3b8cec4b6db5358e07079a88daad124f7b8cca5313b737618e88"} Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.722036 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.722082 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxtlg"] Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.730160 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxtlg"] Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.734416 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.751000 4806 scope.go:117] "RemoveContainer" containerID="ba5872983a56578cc4e17349dedf2fe3b841dbdc0d9bb307e89d4523a23c6a8d" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.766876 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" podStartSLOduration=3.766853056 podStartE2EDuration="3.766853056s" podCreationTimestamp="2025-11-27 10:25:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:25:04.763893691 +0000 UTC m=+209.350484455" watchObservedRunningTime="2025-11-27 10:25:04.766853056 +0000 UTC m=+209.353443820" Nov 27 10:25:04 crc kubenswrapper[4806]: I1127 10:25:04.779459 4806 scope.go:117] "RemoveContainer" containerID="6477e21483913b695296563b76f5d9077c9069c21e5ffb88e8f2936638243d2a" Nov 27 10:25:05 crc kubenswrapper[4806]: I1127 10:25:05.730601 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"38a72bf5-a9ca-409a-b52d-b10ebefc316a","Type":"ContainerStarted","Data":"3ce58a2a37537acf3a2b33025573e96d72e0ce8f4070481d5b5cf22d13f601aa"} Nov 27 10:25:05 crc kubenswrapper[4806]: I1127 10:25:05.733538 4806 generic.go:334] "Generic (PLEG): container finished" podID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" containerID="868dfbdf3ed375064d08ec9331cd332b20c5938c07094c0e32e931f9bf5f1646" exitCode=0 Nov 27 10:25:05 crc kubenswrapper[4806]: I1127 10:25:05.733574 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfw97" event={"ID":"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2","Type":"ContainerDied","Data":"868dfbdf3ed375064d08ec9331cd332b20c5938c07094c0e32e931f9bf5f1646"} Nov 27 10:25:05 crc kubenswrapper[4806]: I1127 10:25:05.739317 4806 generic.go:334] "Generic (PLEG): container finished" podID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" containerID="6d847aa087b075d1492241d80b55736363187ba70efda8d87c94110a62a2c316" exitCode=0 Nov 27 10:25:05 crc kubenswrapper[4806]: I1127 10:25:05.739382 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4x5q" event={"ID":"8cb0392a-70a2-48c0-a69e-4421dbefe86e","Type":"ContainerDied","Data":"6d847aa087b075d1492241d80b55736363187ba70efda8d87c94110a62a2c316"} Nov 27 10:25:05 crc kubenswrapper[4806]: I1127 10:25:05.743143 4806 generic.go:334] "Generic (PLEG): container finished" podID="3209cb7e-221c-4c6a-9864-8101e8171078" containerID="ddcb136b375f253c58a463268b9106aceb9a47d857a54e3c3a790116ee1018e0" exitCode=0 Nov 27 10:25:05 crc kubenswrapper[4806]: I1127 10:25:05.743930 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp6rt" event={"ID":"3209cb7e-221c-4c6a-9864-8101e8171078","Type":"ContainerDied","Data":"ddcb136b375f253c58a463268b9106aceb9a47d857a54e3c3a790116ee1018e0"} Nov 27 10:25:05 crc kubenswrapper[4806]: I1127 10:25:05.755181 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.7551559169999997 podStartE2EDuration="2.755155917s" podCreationTimestamp="2025-11-27 10:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:25:05.752967645 +0000 UTC m=+210.339558409" watchObservedRunningTime="2025-11-27 10:25:05.755155917 +0000 UTC m=+210.341746681" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.128823 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72810aad-3a98-4ed6-bf65-6434e72770fb" path="/var/lib/kubelet/pods/72810aad-3a98-4ed6-bf65-6434e72770fb/volumes" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.288484 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd"] Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.289369 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.294328 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.294719 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.294894 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.295291 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.295324 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.296528 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.306202 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd"] Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.381264 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fb1af7f-2191-420c-a326-aab268e2e9f9-config\") pod \"route-controller-manager-865cf5476-ff2cd\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.381373 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fb1af7f-2191-420c-a326-aab268e2e9f9-serving-cert\") pod \"route-controller-manager-865cf5476-ff2cd\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.381465 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jstv\" (UniqueName: \"kubernetes.io/projected/8fb1af7f-2191-420c-a326-aab268e2e9f9-kube-api-access-5jstv\") pod \"route-controller-manager-865cf5476-ff2cd\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.381585 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fb1af7f-2191-420c-a326-aab268e2e9f9-client-ca\") pod \"route-controller-manager-865cf5476-ff2cd\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.485810 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fb1af7f-2191-420c-a326-aab268e2e9f9-config\") pod \"route-controller-manager-865cf5476-ff2cd\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.485905 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fb1af7f-2191-420c-a326-aab268e2e9f9-serving-cert\") pod \"route-controller-manager-865cf5476-ff2cd\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.485960 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jstv\" (UniqueName: \"kubernetes.io/projected/8fb1af7f-2191-420c-a326-aab268e2e9f9-kube-api-access-5jstv\") pod \"route-controller-manager-865cf5476-ff2cd\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.486007 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fb1af7f-2191-420c-a326-aab268e2e9f9-client-ca\") pod \"route-controller-manager-865cf5476-ff2cd\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.487483 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fb1af7f-2191-420c-a326-aab268e2e9f9-client-ca\") pod \"route-controller-manager-865cf5476-ff2cd\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.494711 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fb1af7f-2191-420c-a326-aab268e2e9f9-config\") pod \"route-controller-manager-865cf5476-ff2cd\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.495357 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fb1af7f-2191-420c-a326-aab268e2e9f9-serving-cert\") pod \"route-controller-manager-865cf5476-ff2cd\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.512992 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jstv\" (UniqueName: \"kubernetes.io/projected/8fb1af7f-2191-420c-a326-aab268e2e9f9-kube-api-access-5jstv\") pod \"route-controller-manager-865cf5476-ff2cd\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.613773 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.772443 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4x5q" event={"ID":"8cb0392a-70a2-48c0-a69e-4421dbefe86e","Type":"ContainerStarted","Data":"7a9678e2f8c22c92b07cff5f423e57a16447e127e36ed2ed5ef3a34e9d03de68"} Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.780521 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp6rt" event={"ID":"3209cb7e-221c-4c6a-9864-8101e8171078","Type":"ContainerStarted","Data":"ab61192ff5176ca7c0370bdd732ef6e16744dd6fc5f396444c78efe04c32275a"} Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.794530 4806 generic.go:334] "Generic (PLEG): container finished" podID="ae461a45-1190-46fc-acf5-8954b951b25e" containerID="0e89b3bc54aa033d4162f98a4172ee5f50be447c6cf2e30add3f877f3b67bc07" exitCode=0 Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.794612 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cthc" event={"ID":"ae461a45-1190-46fc-acf5-8954b951b25e","Type":"ContainerDied","Data":"0e89b3bc54aa033d4162f98a4172ee5f50be447c6cf2e30add3f877f3b67bc07"} Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.809342 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfw97" event={"ID":"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2","Type":"ContainerStarted","Data":"21415667b92d8ac1cbcc776bce593aabfe3e50bd2f18f6b58cea510fdc28713b"} Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.825460 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l4x5q" podStartSLOduration=3.614429338 podStartE2EDuration="56.825434247s" podCreationTimestamp="2025-11-27 10:24:10 +0000 UTC" firstStartedPulling="2025-11-27 10:24:13.087894834 +0000 UTC m=+157.674485598" lastFinishedPulling="2025-11-27 10:25:06.298899743 +0000 UTC m=+210.885490507" observedRunningTime="2025-11-27 10:25:06.803790356 +0000 UTC m=+211.390381140" watchObservedRunningTime="2025-11-27 10:25:06.825434247 +0000 UTC m=+211.412025011" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.827415 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gp6rt" podStartSLOduration=4.5984323929999995 podStartE2EDuration="57.827408713s" podCreationTimestamp="2025-11-27 10:24:09 +0000 UTC" firstStartedPulling="2025-11-27 10:24:13.095559578 +0000 UTC m=+157.682150342" lastFinishedPulling="2025-11-27 10:25:06.324535908 +0000 UTC m=+210.911126662" observedRunningTime="2025-11-27 10:25:06.820583487 +0000 UTC m=+211.407174261" watchObservedRunningTime="2025-11-27 10:25:06.827408713 +0000 UTC m=+211.413999477" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.917704 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wfw97" podStartSLOduration=4.40760313 podStartE2EDuration="59.917657789s" podCreationTimestamp="2025-11-27 10:24:07 +0000 UTC" firstStartedPulling="2025-11-27 10:24:10.756780425 +0000 UTC m=+155.343371199" lastFinishedPulling="2025-11-27 10:25:06.266835094 +0000 UTC m=+210.853425858" observedRunningTime="2025-11-27 10:25:06.878505387 +0000 UTC m=+211.465096151" watchObservedRunningTime="2025-11-27 10:25:06.917657789 +0000 UTC m=+211.504248573" Nov 27 10:25:06 crc kubenswrapper[4806]: I1127 10:25:06.917901 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd"] Nov 27 10:25:06 crc kubenswrapper[4806]: W1127 10:25:06.928370 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fb1af7f_2191_420c_a326_aab268e2e9f9.slice/crio-03498f160b12dc3cb5f0ca2d17637bc109381bb29ad7b22ed72cbf94ed213426 WatchSource:0}: Error finding container 03498f160b12dc3cb5f0ca2d17637bc109381bb29ad7b22ed72cbf94ed213426: Status 404 returned error can't find the container with id 03498f160b12dc3cb5f0ca2d17637bc109381bb29ad7b22ed72cbf94ed213426 Nov 27 10:25:07 crc kubenswrapper[4806]: I1127 10:25:07.103872 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:25:07 crc kubenswrapper[4806]: I1127 10:25:07.155193 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:25:07 crc kubenswrapper[4806]: I1127 10:25:07.818796 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" event={"ID":"8fb1af7f-2191-420c-a326-aab268e2e9f9","Type":"ContainerStarted","Data":"82d18366102e1e1ab687cea857b34a4f041de341a7c521774e651790faa17303"} Nov 27 10:25:07 crc kubenswrapper[4806]: I1127 10:25:07.818860 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" event={"ID":"8fb1af7f-2191-420c-a326-aab268e2e9f9","Type":"ContainerStarted","Data":"03498f160b12dc3cb5f0ca2d17637bc109381bb29ad7b22ed72cbf94ed213426"} Nov 27 10:25:07 crc kubenswrapper[4806]: I1127 10:25:07.818964 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:07 crc kubenswrapper[4806]: I1127 10:25:07.822888 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cthc" event={"ID":"ae461a45-1190-46fc-acf5-8954b951b25e","Type":"ContainerStarted","Data":"94486c2b33041f34e7d95d781468d6251c9c77373525b1e41e5dca082039f561"} Nov 27 10:25:07 crc kubenswrapper[4806]: I1127 10:25:07.833824 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:25:07 crc kubenswrapper[4806]: I1127 10:25:07.841657 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:07 crc kubenswrapper[4806]: I1127 10:25:07.843949 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" podStartSLOduration=6.843931392 podStartE2EDuration="6.843931392s" podCreationTimestamp="2025-11-27 10:25:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:25:07.841890703 +0000 UTC m=+212.428481467" watchObservedRunningTime="2025-11-27 10:25:07.843931392 +0000 UTC m=+212.430522156" Nov 27 10:25:07 crc kubenswrapper[4806]: I1127 10:25:07.914340 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:25:07 crc kubenswrapper[4806]: I1127 10:25:07.947816 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:25:07 crc kubenswrapper[4806]: I1127 10:25:07.947880 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:25:07 crc kubenswrapper[4806]: I1127 10:25:07.962165 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2cthc" podStartSLOduration=4.361936465 podStartE2EDuration="1m0.962140288s" podCreationTimestamp="2025-11-27 10:24:07 +0000 UTC" firstStartedPulling="2025-11-27 10:24:10.687739603 +0000 UTC m=+155.274330357" lastFinishedPulling="2025-11-27 10:25:07.287943426 +0000 UTC m=+211.874534180" observedRunningTime="2025-11-27 10:25:07.958456353 +0000 UTC m=+212.545047117" watchObservedRunningTime="2025-11-27 10:25:07.962140288 +0000 UTC m=+212.548731052" Nov 27 10:25:08 crc kubenswrapper[4806]: I1127 10:25:08.039976 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:25:08 crc kubenswrapper[4806]: I1127 10:25:08.040034 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:25:08 crc kubenswrapper[4806]: I1127 10:25:08.998487 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-wfw97" podUID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" containerName="registry-server" probeResult="failure" output=< Nov 27 10:25:08 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 10:25:08 crc kubenswrapper[4806]: > Nov 27 10:25:09 crc kubenswrapper[4806]: I1127 10:25:09.089706 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-2cthc" podUID="ae461a45-1190-46fc-acf5-8954b951b25e" containerName="registry-server" probeResult="failure" output=< Nov 27 10:25:09 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 10:25:09 crc kubenswrapper[4806]: > Nov 27 10:25:09 crc kubenswrapper[4806]: I1127 10:25:09.657551 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:25:09 crc kubenswrapper[4806]: I1127 10:25:09.657614 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:25:09 crc kubenswrapper[4806]: I1127 10:25:09.701316 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:25:10 crc kubenswrapper[4806]: I1127 10:25:10.507478 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9v4rb"] Nov 27 10:25:10 crc kubenswrapper[4806]: I1127 10:25:10.508185 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9v4rb" podUID="c3c3761a-7b4a-4053-bf59-0634f90d1e91" containerName="registry-server" containerID="cri-o://d6d293d72fab515bfcef36b14063f6a74895a0e88874855d0062a5a2d976c33f" gracePeriod=2 Nov 27 10:25:10 crc kubenswrapper[4806]: I1127 10:25:10.791617 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:25:10 crc kubenswrapper[4806]: I1127 10:25:10.793078 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:25:10 crc kubenswrapper[4806]: I1127 10:25:10.826109 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:25:10 crc kubenswrapper[4806]: I1127 10:25:10.843831 4806 generic.go:334] "Generic (PLEG): container finished" podID="c3c3761a-7b4a-4053-bf59-0634f90d1e91" containerID="d6d293d72fab515bfcef36b14063f6a74895a0e88874855d0062a5a2d976c33f" exitCode=0 Nov 27 10:25:10 crc kubenswrapper[4806]: I1127 10:25:10.844369 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9v4rb" event={"ID":"c3c3761a-7b4a-4053-bf59-0634f90d1e91","Type":"ContainerDied","Data":"d6d293d72fab515bfcef36b14063f6a74895a0e88874855d0062a5a2d976c33f"} Nov 27 10:25:10 crc kubenswrapper[4806]: I1127 10:25:10.876588 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.457138 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.553433 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3c3761a-7b4a-4053-bf59-0634f90d1e91-catalog-content\") pod \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\" (UID: \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\") " Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.553582 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ctb5\" (UniqueName: \"kubernetes.io/projected/c3c3761a-7b4a-4053-bf59-0634f90d1e91-kube-api-access-6ctb5\") pod \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\" (UID: \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\") " Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.553638 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3c3761a-7b4a-4053-bf59-0634f90d1e91-utilities\") pod \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\" (UID: \"c3c3761a-7b4a-4053-bf59-0634f90d1e91\") " Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.554533 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3c3761a-7b4a-4053-bf59-0634f90d1e91-utilities" (OuterVolumeSpecName: "utilities") pod "c3c3761a-7b4a-4053-bf59-0634f90d1e91" (UID: "c3c3761a-7b4a-4053-bf59-0634f90d1e91"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.567506 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3c3761a-7b4a-4053-bf59-0634f90d1e91-kube-api-access-6ctb5" (OuterVolumeSpecName: "kube-api-access-6ctb5") pod "c3c3761a-7b4a-4053-bf59-0634f90d1e91" (UID: "c3c3761a-7b4a-4053-bf59-0634f90d1e91"). InnerVolumeSpecName "kube-api-access-6ctb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.611629 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3c3761a-7b4a-4053-bf59-0634f90d1e91-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3c3761a-7b4a-4053-bf59-0634f90d1e91" (UID: "c3c3761a-7b4a-4053-bf59-0634f90d1e91"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.655178 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ctb5\" (UniqueName: \"kubernetes.io/projected/c3c3761a-7b4a-4053-bf59-0634f90d1e91-kube-api-access-6ctb5\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.655274 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3c3761a-7b4a-4053-bf59-0634f90d1e91-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.655290 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3c3761a-7b4a-4053-bf59-0634f90d1e91-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.676618 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-l956q" podUID="0639fe21-6b99-4fad-be7b-6bd998c3472c" containerName="oauth-openshift" containerID="cri-o://6141b54ffe61baaec35a3bab6aed1a55f0d15f2e5259d6178f1658b6e7522320" gracePeriod=15 Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.829836 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l4x5q" podUID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" containerName="registry-server" probeResult="failure" output=< Nov 27 10:25:11 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 10:25:11 crc kubenswrapper[4806]: > Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.864470 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9v4rb" event={"ID":"c3c3761a-7b4a-4053-bf59-0634f90d1e91","Type":"ContainerDied","Data":"7b8fd3fa27aa38822799a65bae0a1507d81a8aa5ad734e4b97c5ff52228fdbfe"} Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.864593 4806 scope.go:117] "RemoveContainer" containerID="d6d293d72fab515bfcef36b14063f6a74895a0e88874855d0062a5a2d976c33f" Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.864633 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9v4rb" Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.886512 4806 scope.go:117] "RemoveContainer" containerID="8909c19cb3af9cf529da6b2adca42a99919ebf6b1938f3081abdbd49fa9f6af4" Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.912322 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9v4rb"] Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.918097 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9v4rb"] Nov 27 10:25:11 crc kubenswrapper[4806]: I1127 10:25:11.920265 4806 scope.go:117] "RemoveContainer" containerID="405bca87ed37359429e68438556faa68669bfcfdcdf9b7d05474938733777905" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.126084 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3c3761a-7b4a-4053-bf59-0634f90d1e91" path="/var/lib/kubelet/pods/c3c3761a-7b4a-4053-bf59-0634f90d1e91/volumes" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.664746 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.772831 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-router-certs\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.772918 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7qs6\" (UniqueName: \"kubernetes.io/projected/0639fe21-6b99-4fad-be7b-6bd998c3472c-kube-api-access-n7qs6\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.772974 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-audit-policies\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.773005 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-session\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.773041 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0639fe21-6b99-4fad-be7b-6bd998c3472c-audit-dir\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.773070 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-idp-0-file-data\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.773122 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-cliconfig\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.773144 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-provider-selection\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.773183 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-trusted-ca-bundle\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.773210 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-serving-cert\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.773263 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-error\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.774173 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.774203 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.774635 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-service-ca\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.774681 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-ocp-branding-template\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.774721 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-login\") pod \"0639fe21-6b99-4fad-be7b-6bd998c3472c\" (UID: \"0639fe21-6b99-4fad-be7b-6bd998c3472c\") " Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.774780 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0639fe21-6b99-4fad-be7b-6bd998c3472c-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.775604 4806 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0639fe21-6b99-4fad-be7b-6bd998c3472c-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.775630 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.775643 4806 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.775792 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.776605 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.779526 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.779794 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.779930 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.779968 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0639fe21-6b99-4fad-be7b-6bd998c3472c-kube-api-access-n7qs6" (OuterVolumeSpecName: "kube-api-access-n7qs6") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "kube-api-access-n7qs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.780352 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.780510 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.783728 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.784248 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.785438 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "0639fe21-6b99-4fad-be7b-6bd998c3472c" (UID: "0639fe21-6b99-4fad-be7b-6bd998c3472c"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.871942 4806 generic.go:334] "Generic (PLEG): container finished" podID="0639fe21-6b99-4fad-be7b-6bd998c3472c" containerID="6141b54ffe61baaec35a3bab6aed1a55f0d15f2e5259d6178f1658b6e7522320" exitCode=0 Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.872053 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-l956q" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.872343 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l956q" event={"ID":"0639fe21-6b99-4fad-be7b-6bd998c3472c","Type":"ContainerDied","Data":"6141b54ffe61baaec35a3bab6aed1a55f0d15f2e5259d6178f1658b6e7522320"} Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.872493 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-l956q" event={"ID":"0639fe21-6b99-4fad-be7b-6bd998c3472c","Type":"ContainerDied","Data":"017278a159ddfb26c4b4693cccca6595d5987738690262a9d9f06a9f3620f1e6"} Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.872521 4806 scope.go:117] "RemoveContainer" containerID="6141b54ffe61baaec35a3bab6aed1a55f0d15f2e5259d6178f1658b6e7522320" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.877804 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.877871 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7qs6\" (UniqueName: \"kubernetes.io/projected/0639fe21-6b99-4fad-be7b-6bd998c3472c-kube-api-access-n7qs6\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.877882 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.877892 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.877904 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.877914 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.877926 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.877937 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.877946 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.877957 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.877967 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0639fe21-6b99-4fad-be7b-6bd998c3472c-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.901063 4806 scope.go:117] "RemoveContainer" containerID="6141b54ffe61baaec35a3bab6aed1a55f0d15f2e5259d6178f1658b6e7522320" Nov 27 10:25:12 crc kubenswrapper[4806]: E1127 10:25:12.901839 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6141b54ffe61baaec35a3bab6aed1a55f0d15f2e5259d6178f1658b6e7522320\": container with ID starting with 6141b54ffe61baaec35a3bab6aed1a55f0d15f2e5259d6178f1658b6e7522320 not found: ID does not exist" containerID="6141b54ffe61baaec35a3bab6aed1a55f0d15f2e5259d6178f1658b6e7522320" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.901892 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6141b54ffe61baaec35a3bab6aed1a55f0d15f2e5259d6178f1658b6e7522320"} err="failed to get container status \"6141b54ffe61baaec35a3bab6aed1a55f0d15f2e5259d6178f1658b6e7522320\": rpc error: code = NotFound desc = could not find container \"6141b54ffe61baaec35a3bab6aed1a55f0d15f2e5259d6178f1658b6e7522320\": container with ID starting with 6141b54ffe61baaec35a3bab6aed1a55f0d15f2e5259d6178f1658b6e7522320 not found: ID does not exist" Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.905315 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l956q"] Nov 27 10:25:12 crc kubenswrapper[4806]: I1127 10:25:12.907755 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-l956q"] Nov 27 10:25:14 crc kubenswrapper[4806]: I1127 10:25:14.131187 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0639fe21-6b99-4fad-be7b-6bd998c3472c" path="/var/lib/kubelet/pods/0639fe21-6b99-4fad-be7b-6bd998c3472c/volumes" Nov 27 10:25:14 crc kubenswrapper[4806]: I1127 10:25:14.450596 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:25:14 crc kubenswrapper[4806]: I1127 10:25:14.450708 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:25:14 crc kubenswrapper[4806]: I1127 10:25:14.450786 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:25:14 crc kubenswrapper[4806]: I1127 10:25:14.451742 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218"} pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 10:25:14 crc kubenswrapper[4806]: I1127 10:25:14.451818 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" containerID="cri-o://7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218" gracePeriod=600 Nov 27 10:25:14 crc kubenswrapper[4806]: I1127 10:25:14.891035 4806 generic.go:334] "Generic (PLEG): container finished" podID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerID="7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218" exitCode=0 Nov 27 10:25:14 crc kubenswrapper[4806]: I1127 10:25:14.891116 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerDied","Data":"7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218"} Nov 27 10:25:14 crc kubenswrapper[4806]: I1127 10:25:14.891168 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"0f2788d6ded82e196b1f288174c14861105e5b2e48bf181245fe8449fb3b1acb"} Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.297035 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6fb796c88-wgqmj"] Nov 27 10:25:16 crc kubenswrapper[4806]: E1127 10:25:16.299363 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c3761a-7b4a-4053-bf59-0634f90d1e91" containerName="extract-utilities" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.299467 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c3761a-7b4a-4053-bf59-0634f90d1e91" containerName="extract-utilities" Nov 27 10:25:16 crc kubenswrapper[4806]: E1127 10:25:16.299534 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c3761a-7b4a-4053-bf59-0634f90d1e91" containerName="registry-server" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.299596 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c3761a-7b4a-4053-bf59-0634f90d1e91" containerName="registry-server" Nov 27 10:25:16 crc kubenswrapper[4806]: E1127 10:25:16.299667 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c3761a-7b4a-4053-bf59-0634f90d1e91" containerName="extract-content" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.299725 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c3761a-7b4a-4053-bf59-0634f90d1e91" containerName="extract-content" Nov 27 10:25:16 crc kubenswrapper[4806]: E1127 10:25:16.299787 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0639fe21-6b99-4fad-be7b-6bd998c3472c" containerName="oauth-openshift" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.299850 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="0639fe21-6b99-4fad-be7b-6bd998c3472c" containerName="oauth-openshift" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.300030 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="0639fe21-6b99-4fad-be7b-6bd998c3472c" containerName="oauth-openshift" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.300108 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3c3761a-7b4a-4053-bf59-0634f90d1e91" containerName="registry-server" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.300629 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.304905 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.305013 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.305145 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.305180 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.305287 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.304905 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.306972 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.307140 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.307583 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.307743 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.308026 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.346989 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0d9b1918-bf62-46a5-bbe1-5666822166ee-audit-policies\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.347041 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.347063 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.347084 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.347110 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.347126 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0d9b1918-bf62-46a5-bbe1-5666822166ee-audit-dir\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.347148 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-router-certs\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.347170 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.347190 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-service-ca\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.347209 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.347245 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-session\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.347304 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-user-template-error\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.347329 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-user-template-login\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.347396 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kkwt\" (UniqueName: \"kubernetes.io/projected/0d9b1918-bf62-46a5-bbe1-5666822166ee-kube-api-access-2kkwt\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.350471 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.350678 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.352460 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6fb796c88-wgqmj"] Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.352589 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.359923 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448475 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-user-template-error\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448547 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-user-template-login\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448579 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kkwt\" (UniqueName: \"kubernetes.io/projected/0d9b1918-bf62-46a5-bbe1-5666822166ee-kube-api-access-2kkwt\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448614 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0d9b1918-bf62-46a5-bbe1-5666822166ee-audit-policies\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448648 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448674 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448701 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448731 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448756 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0d9b1918-bf62-46a5-bbe1-5666822166ee-audit-dir\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448786 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-router-certs\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448814 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448844 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-service-ca\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448867 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.448898 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-session\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.450329 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0d9b1918-bf62-46a5-bbe1-5666822166ee-audit-dir\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.450789 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-service-ca\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.450943 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0d9b1918-bf62-46a5-bbe1-5666822166ee-audit-policies\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.452664 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.453208 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.456138 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-session\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.456169 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.456478 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-user-template-error\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.456793 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.457127 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.459013 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.460381 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-system-router-certs\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.465286 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0d9b1918-bf62-46a5-bbe1-5666822166ee-v4-0-config-user-template-login\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.468071 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kkwt\" (UniqueName: \"kubernetes.io/projected/0d9b1918-bf62-46a5-bbe1-5666822166ee-kube-api-access-2kkwt\") pod \"oauth-openshift-6fb796c88-wgqmj\" (UID: \"0d9b1918-bf62-46a5-bbe1-5666822166ee\") " pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:16 crc kubenswrapper[4806]: I1127 10:25:16.672020 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:17 crc kubenswrapper[4806]: I1127 10:25:17.103817 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6fb796c88-wgqmj"] Nov 27 10:25:17 crc kubenswrapper[4806]: I1127 10:25:17.919358 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" event={"ID":"0d9b1918-bf62-46a5-bbe1-5666822166ee","Type":"ContainerStarted","Data":"b12869172f3fcd96dadfb9f3325f49b323f4827f8a0b373b4391dea7345e0d1f"} Nov 27 10:25:17 crc kubenswrapper[4806]: I1127 10:25:17.920869 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" event={"ID":"0d9b1918-bf62-46a5-bbe1-5666822166ee","Type":"ContainerStarted","Data":"847cc3424931d19329749eb96ac4a372422beddffec1ac0c612128ce948a944f"} Nov 27 10:25:17 crc kubenswrapper[4806]: I1127 10:25:17.920898 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:17 crc kubenswrapper[4806]: I1127 10:25:17.927837 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" Nov 27 10:25:17 crc kubenswrapper[4806]: I1127 10:25:17.951421 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6fb796c88-wgqmj" podStartSLOduration=31.951393351 podStartE2EDuration="31.951393351s" podCreationTimestamp="2025-11-27 10:24:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:25:17.945598415 +0000 UTC m=+222.532189189" watchObservedRunningTime="2025-11-27 10:25:17.951393351 +0000 UTC m=+222.537984115" Nov 27 10:25:18 crc kubenswrapper[4806]: I1127 10:25:18.025302 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:25:18 crc kubenswrapper[4806]: I1127 10:25:18.111546 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:25:18 crc kubenswrapper[4806]: I1127 10:25:18.132206 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:25:18 crc kubenswrapper[4806]: I1127 10:25:18.180353 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:25:19 crc kubenswrapper[4806]: I1127 10:25:19.281525 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2cthc"] Nov 27 10:25:19 crc kubenswrapper[4806]: I1127 10:25:19.699773 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:25:19 crc kubenswrapper[4806]: I1127 10:25:19.929709 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2cthc" podUID="ae461a45-1190-46fc-acf5-8954b951b25e" containerName="registry-server" containerID="cri-o://94486c2b33041f34e7d95d781468d6251c9c77373525b1e41e5dca082039f561" gracePeriod=2 Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.411982 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.414056 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae461a45-1190-46fc-acf5-8954b951b25e-utilities\") pod \"ae461a45-1190-46fc-acf5-8954b951b25e\" (UID: \"ae461a45-1190-46fc-acf5-8954b951b25e\") " Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.414102 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae461a45-1190-46fc-acf5-8954b951b25e-catalog-content\") pod \"ae461a45-1190-46fc-acf5-8954b951b25e\" (UID: \"ae461a45-1190-46fc-acf5-8954b951b25e\") " Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.414161 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mz5t\" (UniqueName: \"kubernetes.io/projected/ae461a45-1190-46fc-acf5-8954b951b25e-kube-api-access-7mz5t\") pod \"ae461a45-1190-46fc-acf5-8954b951b25e\" (UID: \"ae461a45-1190-46fc-acf5-8954b951b25e\") " Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.415089 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae461a45-1190-46fc-acf5-8954b951b25e-utilities" (OuterVolumeSpecName: "utilities") pod "ae461a45-1190-46fc-acf5-8954b951b25e" (UID: "ae461a45-1190-46fc-acf5-8954b951b25e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.421599 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae461a45-1190-46fc-acf5-8954b951b25e-kube-api-access-7mz5t" (OuterVolumeSpecName: "kube-api-access-7mz5t") pod "ae461a45-1190-46fc-acf5-8954b951b25e" (UID: "ae461a45-1190-46fc-acf5-8954b951b25e"). InnerVolumeSpecName "kube-api-access-7mz5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.468890 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae461a45-1190-46fc-acf5-8954b951b25e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae461a45-1190-46fc-acf5-8954b951b25e" (UID: "ae461a45-1190-46fc-acf5-8954b951b25e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.515364 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae461a45-1190-46fc-acf5-8954b951b25e-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.515409 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae461a45-1190-46fc-acf5-8954b951b25e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.515423 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mz5t\" (UniqueName: \"kubernetes.io/projected/ae461a45-1190-46fc-acf5-8954b951b25e-kube-api-access-7mz5t\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.840574 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.887403 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.937353 4806 generic.go:334] "Generic (PLEG): container finished" podID="ae461a45-1190-46fc-acf5-8954b951b25e" containerID="94486c2b33041f34e7d95d781468d6251c9c77373525b1e41e5dca082039f561" exitCode=0 Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.937435 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cthc" Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.937509 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cthc" event={"ID":"ae461a45-1190-46fc-acf5-8954b951b25e","Type":"ContainerDied","Data":"94486c2b33041f34e7d95d781468d6251c9c77373525b1e41e5dca082039f561"} Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.937549 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cthc" event={"ID":"ae461a45-1190-46fc-acf5-8954b951b25e","Type":"ContainerDied","Data":"f50a9bb1d7b1ceb7a6370bfff88a3b3d8e38d7328e2c39a7fbaf697bcc3ad46b"} Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.937575 4806 scope.go:117] "RemoveContainer" containerID="94486c2b33041f34e7d95d781468d6251c9c77373525b1e41e5dca082039f561" Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.959936 4806 scope.go:117] "RemoveContainer" containerID="0e89b3bc54aa033d4162f98a4172ee5f50be447c6cf2e30add3f877f3b67bc07" Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.983870 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2cthc"] Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.993034 4806 scope.go:117] "RemoveContainer" containerID="31f320c341b64c5a2dadedd741fe3c4ad5719b3cb977c1ddb38df188b200a04b" Nov 27 10:25:20 crc kubenswrapper[4806]: I1127 10:25:20.993166 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2cthc"] Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.007792 4806 scope.go:117] "RemoveContainer" containerID="94486c2b33041f34e7d95d781468d6251c9c77373525b1e41e5dca082039f561" Nov 27 10:25:21 crc kubenswrapper[4806]: E1127 10:25:21.008994 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94486c2b33041f34e7d95d781468d6251c9c77373525b1e41e5dca082039f561\": container with ID starting with 94486c2b33041f34e7d95d781468d6251c9c77373525b1e41e5dca082039f561 not found: ID does not exist" containerID="94486c2b33041f34e7d95d781468d6251c9c77373525b1e41e5dca082039f561" Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.009058 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94486c2b33041f34e7d95d781468d6251c9c77373525b1e41e5dca082039f561"} err="failed to get container status \"94486c2b33041f34e7d95d781468d6251c9c77373525b1e41e5dca082039f561\": rpc error: code = NotFound desc = could not find container \"94486c2b33041f34e7d95d781468d6251c9c77373525b1e41e5dca082039f561\": container with ID starting with 94486c2b33041f34e7d95d781468d6251c9c77373525b1e41e5dca082039f561 not found: ID does not exist" Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.009099 4806 scope.go:117] "RemoveContainer" containerID="0e89b3bc54aa033d4162f98a4172ee5f50be447c6cf2e30add3f877f3b67bc07" Nov 27 10:25:21 crc kubenswrapper[4806]: E1127 10:25:21.009495 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e89b3bc54aa033d4162f98a4172ee5f50be447c6cf2e30add3f877f3b67bc07\": container with ID starting with 0e89b3bc54aa033d4162f98a4172ee5f50be447c6cf2e30add3f877f3b67bc07 not found: ID does not exist" containerID="0e89b3bc54aa033d4162f98a4172ee5f50be447c6cf2e30add3f877f3b67bc07" Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.009526 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e89b3bc54aa033d4162f98a4172ee5f50be447c6cf2e30add3f877f3b67bc07"} err="failed to get container status \"0e89b3bc54aa033d4162f98a4172ee5f50be447c6cf2e30add3f877f3b67bc07\": rpc error: code = NotFound desc = could not find container \"0e89b3bc54aa033d4162f98a4172ee5f50be447c6cf2e30add3f877f3b67bc07\": container with ID starting with 0e89b3bc54aa033d4162f98a4172ee5f50be447c6cf2e30add3f877f3b67bc07 not found: ID does not exist" Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.009547 4806 scope.go:117] "RemoveContainer" containerID="31f320c341b64c5a2dadedd741fe3c4ad5719b3cb977c1ddb38df188b200a04b" Nov 27 10:25:21 crc kubenswrapper[4806]: E1127 10:25:21.010814 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31f320c341b64c5a2dadedd741fe3c4ad5719b3cb977c1ddb38df188b200a04b\": container with ID starting with 31f320c341b64c5a2dadedd741fe3c4ad5719b3cb977c1ddb38df188b200a04b not found: ID does not exist" containerID="31f320c341b64c5a2dadedd741fe3c4ad5719b3cb977c1ddb38df188b200a04b" Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.010871 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31f320c341b64c5a2dadedd741fe3c4ad5719b3cb977c1ddb38df188b200a04b"} err="failed to get container status \"31f320c341b64c5a2dadedd741fe3c4ad5719b3cb977c1ddb38df188b200a04b\": rpc error: code = NotFound desc = could not find container \"31f320c341b64c5a2dadedd741fe3c4ad5719b3cb977c1ddb38df188b200a04b\": container with ID starting with 31f320c341b64c5a2dadedd741fe3c4ad5719b3cb977c1ddb38df188b200a04b not found: ID does not exist" Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.318028 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-55568dc997-pmlhd"] Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.318574 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" podUID="d09a5847-2755-42b2-a220-a3b254b3ba19" containerName="controller-manager" containerID="cri-o://9fd2be668fa8835cdd4446a034366d9183ad716747dab93d08d3aa690cd9ea15" gracePeriod=30 Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.347149 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd"] Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.347423 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" podUID="8fb1af7f-2191-420c-a326-aab268e2e9f9" containerName="route-controller-manager" containerID="cri-o://82d18366102e1e1ab687cea857b34a4f041de341a7c521774e651790faa17303" gracePeriod=30 Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.943929 4806 generic.go:334] "Generic (PLEG): container finished" podID="8fb1af7f-2191-420c-a326-aab268e2e9f9" containerID="82d18366102e1e1ab687cea857b34a4f041de341a7c521774e651790faa17303" exitCode=0 Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.944015 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" event={"ID":"8fb1af7f-2191-420c-a326-aab268e2e9f9","Type":"ContainerDied","Data":"82d18366102e1e1ab687cea857b34a4f041de341a7c521774e651790faa17303"} Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.947443 4806 generic.go:334] "Generic (PLEG): container finished" podID="d09a5847-2755-42b2-a220-a3b254b3ba19" containerID="9fd2be668fa8835cdd4446a034366d9183ad716747dab93d08d3aa690cd9ea15" exitCode=0 Nov 27 10:25:21 crc kubenswrapper[4806]: I1127 10:25:21.947489 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" event={"ID":"d09a5847-2755-42b2-a220-a3b254b3ba19","Type":"ContainerDied","Data":"9fd2be668fa8835cdd4446a034366d9183ad716747dab93d08d3aa690cd9ea15"} Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.126469 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae461a45-1190-46fc-acf5-8954b951b25e" path="/var/lib/kubelet/pods/ae461a45-1190-46fc-acf5-8954b951b25e/volumes" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.377491 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.384452 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.404592 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-c659fc556-5s7t7"] Nov 27 10:25:22 crc kubenswrapper[4806]: E1127 10:25:22.404907 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae461a45-1190-46fc-acf5-8954b951b25e" containerName="extract-content" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.404926 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae461a45-1190-46fc-acf5-8954b951b25e" containerName="extract-content" Nov 27 10:25:22 crc kubenswrapper[4806]: E1127 10:25:22.404936 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae461a45-1190-46fc-acf5-8954b951b25e" containerName="registry-server" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.404943 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae461a45-1190-46fc-acf5-8954b951b25e" containerName="registry-server" Nov 27 10:25:22 crc kubenswrapper[4806]: E1127 10:25:22.404965 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae461a45-1190-46fc-acf5-8954b951b25e" containerName="extract-utilities" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.404975 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae461a45-1190-46fc-acf5-8954b951b25e" containerName="extract-utilities" Nov 27 10:25:22 crc kubenswrapper[4806]: E1127 10:25:22.404989 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d09a5847-2755-42b2-a220-a3b254b3ba19" containerName="controller-manager" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.404997 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d09a5847-2755-42b2-a220-a3b254b3ba19" containerName="controller-manager" Nov 27 10:25:22 crc kubenswrapper[4806]: E1127 10:25:22.405008 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fb1af7f-2191-420c-a326-aab268e2e9f9" containerName="route-controller-manager" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.405018 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fb1af7f-2191-420c-a326-aab268e2e9f9" containerName="route-controller-manager" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.405161 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d09a5847-2755-42b2-a220-a3b254b3ba19" containerName="controller-manager" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.405177 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fb1af7f-2191-420c-a326-aab268e2e9f9" containerName="route-controller-manager" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.405193 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae461a45-1190-46fc-acf5-8954b951b25e" containerName="registry-server" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.405856 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.423792 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c659fc556-5s7t7"] Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543122 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-client-ca\") pod \"d09a5847-2755-42b2-a220-a3b254b3ba19\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543220 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-config\") pod \"d09a5847-2755-42b2-a220-a3b254b3ba19\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543301 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n848\" (UniqueName: \"kubernetes.io/projected/d09a5847-2755-42b2-a220-a3b254b3ba19-kube-api-access-2n848\") pod \"d09a5847-2755-42b2-a220-a3b254b3ba19\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543331 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d09a5847-2755-42b2-a220-a3b254b3ba19-serving-cert\") pod \"d09a5847-2755-42b2-a220-a3b254b3ba19\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543398 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fb1af7f-2191-420c-a326-aab268e2e9f9-client-ca\") pod \"8fb1af7f-2191-420c-a326-aab268e2e9f9\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543434 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-proxy-ca-bundles\") pod \"d09a5847-2755-42b2-a220-a3b254b3ba19\" (UID: \"d09a5847-2755-42b2-a220-a3b254b3ba19\") " Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543464 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fb1af7f-2191-420c-a326-aab268e2e9f9-serving-cert\") pod \"8fb1af7f-2191-420c-a326-aab268e2e9f9\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543482 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jstv\" (UniqueName: \"kubernetes.io/projected/8fb1af7f-2191-420c-a326-aab268e2e9f9-kube-api-access-5jstv\") pod \"8fb1af7f-2191-420c-a326-aab268e2e9f9\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543519 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fb1af7f-2191-420c-a326-aab268e2e9f9-config\") pod \"8fb1af7f-2191-420c-a326-aab268e2e9f9\" (UID: \"8fb1af7f-2191-420c-a326-aab268e2e9f9\") " Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543733 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-serving-cert\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543776 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-proxy-ca-bundles\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543806 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-config\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543833 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxvjj\" (UniqueName: \"kubernetes.io/projected/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-kube-api-access-rxvjj\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.543851 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-client-ca\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.544659 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d09a5847-2755-42b2-a220-a3b254b3ba19" (UID: "d09a5847-2755-42b2-a220-a3b254b3ba19"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.544663 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fb1af7f-2191-420c-a326-aab268e2e9f9-client-ca" (OuterVolumeSpecName: "client-ca") pod "8fb1af7f-2191-420c-a326-aab268e2e9f9" (UID: "8fb1af7f-2191-420c-a326-aab268e2e9f9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.545158 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-client-ca" (OuterVolumeSpecName: "client-ca") pod "d09a5847-2755-42b2-a220-a3b254b3ba19" (UID: "d09a5847-2755-42b2-a220-a3b254b3ba19"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.545387 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fb1af7f-2191-420c-a326-aab268e2e9f9-config" (OuterVolumeSpecName: "config") pod "8fb1af7f-2191-420c-a326-aab268e2e9f9" (UID: "8fb1af7f-2191-420c-a326-aab268e2e9f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.546208 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-config" (OuterVolumeSpecName: "config") pod "d09a5847-2755-42b2-a220-a3b254b3ba19" (UID: "d09a5847-2755-42b2-a220-a3b254b3ba19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.550351 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fb1af7f-2191-420c-a326-aab268e2e9f9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8fb1af7f-2191-420c-a326-aab268e2e9f9" (UID: "8fb1af7f-2191-420c-a326-aab268e2e9f9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.550732 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d09a5847-2755-42b2-a220-a3b254b3ba19-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d09a5847-2755-42b2-a220-a3b254b3ba19" (UID: "d09a5847-2755-42b2-a220-a3b254b3ba19"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.551420 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d09a5847-2755-42b2-a220-a3b254b3ba19-kube-api-access-2n848" (OuterVolumeSpecName: "kube-api-access-2n848") pod "d09a5847-2755-42b2-a220-a3b254b3ba19" (UID: "d09a5847-2755-42b2-a220-a3b254b3ba19"). InnerVolumeSpecName "kube-api-access-2n848". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.553349 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fb1af7f-2191-420c-a326-aab268e2e9f9-kube-api-access-5jstv" (OuterVolumeSpecName: "kube-api-access-5jstv") pod "8fb1af7f-2191-420c-a326-aab268e2e9f9" (UID: "8fb1af7f-2191-420c-a326-aab268e2e9f9"). InnerVolumeSpecName "kube-api-access-5jstv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645381 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-serving-cert\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645461 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-proxy-ca-bundles\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645500 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-config\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645533 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxvjj\" (UniqueName: \"kubernetes.io/projected/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-kube-api-access-rxvjj\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645557 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-client-ca\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645624 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645639 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n848\" (UniqueName: \"kubernetes.io/projected/d09a5847-2755-42b2-a220-a3b254b3ba19-kube-api-access-2n848\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645650 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d09a5847-2755-42b2-a220-a3b254b3ba19-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645662 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fb1af7f-2191-420c-a326-aab268e2e9f9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645671 4806 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645681 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fb1af7f-2191-420c-a326-aab268e2e9f9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645691 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jstv\" (UniqueName: \"kubernetes.io/projected/8fb1af7f-2191-420c-a326-aab268e2e9f9-kube-api-access-5jstv\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645701 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fb1af7f-2191-420c-a326-aab268e2e9f9-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.645710 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d09a5847-2755-42b2-a220-a3b254b3ba19-client-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.646912 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-client-ca\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.647764 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-config\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.648499 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-proxy-ca-bundles\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.658471 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-serving-cert\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.661930 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxvjj\" (UniqueName: \"kubernetes.io/projected/29be093f-5ac6-4ca5-8d30-3e8fe2d1903c-kube-api-access-rxvjj\") pod \"controller-manager-c659fc556-5s7t7\" (UID: \"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c\") " pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.738103 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.956460 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.956460 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-55568dc997-pmlhd" event={"ID":"d09a5847-2755-42b2-a220-a3b254b3ba19","Type":"ContainerDied","Data":"56c3f323e19e3b8cec4b6db5358e07079a88daad124f7b8cca5313b737618e88"} Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.956625 4806 scope.go:117] "RemoveContainer" containerID="9fd2be668fa8835cdd4446a034366d9183ad716747dab93d08d3aa690cd9ea15" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.958190 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" event={"ID":"8fb1af7f-2191-420c-a326-aab268e2e9f9","Type":"ContainerDied","Data":"03498f160b12dc3cb5f0ca2d17637bc109381bb29ad7b22ed72cbf94ed213426"} Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.958279 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.977636 4806 scope.go:117] "RemoveContainer" containerID="82d18366102e1e1ab687cea857b34a4f041de341a7c521774e651790faa17303" Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.990191 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-55568dc997-pmlhd"] Nov 27 10:25:22 crc kubenswrapper[4806]: I1127 10:25:22.994307 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-55568dc997-pmlhd"] Nov 27 10:25:23 crc kubenswrapper[4806]: I1127 10:25:23.004353 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd"] Nov 27 10:25:23 crc kubenswrapper[4806]: I1127 10:25:23.016182 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-865cf5476-ff2cd"] Nov 27 10:25:23 crc kubenswrapper[4806]: I1127 10:25:23.134125 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c659fc556-5s7t7"] Nov 27 10:25:23 crc kubenswrapper[4806]: W1127 10:25:23.140473 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29be093f_5ac6_4ca5_8d30_3e8fe2d1903c.slice/crio-d067b26da031ed6cd97acdaae22302b294244ac423f2d7e2c7def5736a75c502 WatchSource:0}: Error finding container d067b26da031ed6cd97acdaae22302b294244ac423f2d7e2c7def5736a75c502: Status 404 returned error can't find the container with id d067b26da031ed6cd97acdaae22302b294244ac423f2d7e2c7def5736a75c502 Nov 27 10:25:23 crc kubenswrapper[4806]: I1127 10:25:23.682058 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l4x5q"] Nov 27 10:25:23 crc kubenswrapper[4806]: I1127 10:25:23.682846 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l4x5q" podUID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" containerName="registry-server" containerID="cri-o://7a9678e2f8c22c92b07cff5f423e57a16447e127e36ed2ed5ef3a34e9d03de68" gracePeriod=2 Nov 27 10:25:23 crc kubenswrapper[4806]: I1127 10:25:23.970933 4806 generic.go:334] "Generic (PLEG): container finished" podID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" containerID="7a9678e2f8c22c92b07cff5f423e57a16447e127e36ed2ed5ef3a34e9d03de68" exitCode=0 Nov 27 10:25:23 crc kubenswrapper[4806]: I1127 10:25:23.971004 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4x5q" event={"ID":"8cb0392a-70a2-48c0-a69e-4421dbefe86e","Type":"ContainerDied","Data":"7a9678e2f8c22c92b07cff5f423e57a16447e127e36ed2ed5ef3a34e9d03de68"} Nov 27 10:25:23 crc kubenswrapper[4806]: I1127 10:25:23.974270 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" event={"ID":"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c","Type":"ContainerStarted","Data":"66a91fe697c659403723a7f0ea3ba4049a628a747e432d31858210706c4d74fe"} Nov 27 10:25:23 crc kubenswrapper[4806]: I1127 10:25:23.974327 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" event={"ID":"29be093f-5ac6-4ca5-8d30-3e8fe2d1903c","Type":"ContainerStarted","Data":"d067b26da031ed6cd97acdaae22302b294244ac423f2d7e2c7def5736a75c502"} Nov 27 10:25:23 crc kubenswrapper[4806]: I1127 10:25:23.996702 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" podStartSLOduration=2.996675468 podStartE2EDuration="2.996675468s" podCreationTimestamp="2025-11-27 10:25:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:25:23.995457623 +0000 UTC m=+228.582048407" watchObservedRunningTime="2025-11-27 10:25:23.996675468 +0000 UTC m=+228.583266232" Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.046586 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.124831 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fb1af7f-2191-420c-a326-aab268e2e9f9" path="/var/lib/kubelet/pods/8fb1af7f-2191-420c-a326-aab268e2e9f9/volumes" Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.125530 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d09a5847-2755-42b2-a220-a3b254b3ba19" path="/var/lib/kubelet/pods/d09a5847-2755-42b2-a220-a3b254b3ba19/volumes" Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.168891 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cb0392a-70a2-48c0-a69e-4421dbefe86e-catalog-content\") pod \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\" (UID: \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\") " Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.169122 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z57v\" (UniqueName: \"kubernetes.io/projected/8cb0392a-70a2-48c0-a69e-4421dbefe86e-kube-api-access-8z57v\") pod \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\" (UID: \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\") " Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.169153 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cb0392a-70a2-48c0-a69e-4421dbefe86e-utilities\") pod \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\" (UID: \"8cb0392a-70a2-48c0-a69e-4421dbefe86e\") " Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.170052 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cb0392a-70a2-48c0-a69e-4421dbefe86e-utilities" (OuterVolumeSpecName: "utilities") pod "8cb0392a-70a2-48c0-a69e-4421dbefe86e" (UID: "8cb0392a-70a2-48c0-a69e-4421dbefe86e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.177315 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cb0392a-70a2-48c0-a69e-4421dbefe86e-kube-api-access-8z57v" (OuterVolumeSpecName: "kube-api-access-8z57v") pod "8cb0392a-70a2-48c0-a69e-4421dbefe86e" (UID: "8cb0392a-70a2-48c0-a69e-4421dbefe86e"). InnerVolumeSpecName "kube-api-access-8z57v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.271032 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z57v\" (UniqueName: \"kubernetes.io/projected/8cb0392a-70a2-48c0-a69e-4421dbefe86e-kube-api-access-8z57v\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.271105 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cb0392a-70a2-48c0-a69e-4421dbefe86e-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.283930 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cb0392a-70a2-48c0-a69e-4421dbefe86e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8cb0392a-70a2-48c0-a69e-4421dbefe86e" (UID: "8cb0392a-70a2-48c0-a69e-4421dbefe86e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.371954 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cb0392a-70a2-48c0-a69e-4421dbefe86e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.986773 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l4x5q" event={"ID":"8cb0392a-70a2-48c0-a69e-4421dbefe86e","Type":"ContainerDied","Data":"73ea34e538f10958ca6c07b967476ab1f12b959729382fa29bf8ec7de19ccfc3"} Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.986929 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.986815 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l4x5q" Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.986961 4806 scope.go:117] "RemoveContainer" containerID="7a9678e2f8c22c92b07cff5f423e57a16447e127e36ed2ed5ef3a34e9d03de68" Nov 27 10:25:24 crc kubenswrapper[4806]: I1127 10:25:24.993446 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-c659fc556-5s7t7" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.008621 4806 scope.go:117] "RemoveContainer" containerID="6d847aa087b075d1492241d80b55736363187ba70efda8d87c94110a62a2c316" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.036519 4806 scope.go:117] "RemoveContainer" containerID="50c6ff7f5250333d4f238fb01996502c749316c4d66d04c0151d9903dbd3cb5a" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.066579 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l4x5q"] Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.080415 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l4x5q"] Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.312342 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c"] Nov 27 10:25:25 crc kubenswrapper[4806]: E1127 10:25:25.312848 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" containerName="registry-server" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.312927 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" containerName="registry-server" Nov 27 10:25:25 crc kubenswrapper[4806]: E1127 10:25:25.312995 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" containerName="extract-utilities" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.313068 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" containerName="extract-utilities" Nov 27 10:25:25 crc kubenswrapper[4806]: E1127 10:25:25.313152 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" containerName="extract-content" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.313208 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" containerName="extract-content" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.313395 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" containerName="registry-server" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.313863 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.315825 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.316416 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.316660 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.321249 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.321351 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.321521 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c"] Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.322748 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.490995 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97e2decb-97dc-49b8-a506-e9009b1e4af1-config\") pod \"route-controller-manager-6d679bc5f5-b6c7c\" (UID: \"97e2decb-97dc-49b8-a506-e9009b1e4af1\") " pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.491046 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97e2decb-97dc-49b8-a506-e9009b1e4af1-client-ca\") pod \"route-controller-manager-6d679bc5f5-b6c7c\" (UID: \"97e2decb-97dc-49b8-a506-e9009b1e4af1\") " pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.491213 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97e2decb-97dc-49b8-a506-e9009b1e4af1-serving-cert\") pod \"route-controller-manager-6d679bc5f5-b6c7c\" (UID: \"97e2decb-97dc-49b8-a506-e9009b1e4af1\") " pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.491297 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98cth\" (UniqueName: \"kubernetes.io/projected/97e2decb-97dc-49b8-a506-e9009b1e4af1-kube-api-access-98cth\") pod \"route-controller-manager-6d679bc5f5-b6c7c\" (UID: \"97e2decb-97dc-49b8-a506-e9009b1e4af1\") " pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.592137 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97e2decb-97dc-49b8-a506-e9009b1e4af1-config\") pod \"route-controller-manager-6d679bc5f5-b6c7c\" (UID: \"97e2decb-97dc-49b8-a506-e9009b1e4af1\") " pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.592191 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97e2decb-97dc-49b8-a506-e9009b1e4af1-client-ca\") pod \"route-controller-manager-6d679bc5f5-b6c7c\" (UID: \"97e2decb-97dc-49b8-a506-e9009b1e4af1\") " pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.592214 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97e2decb-97dc-49b8-a506-e9009b1e4af1-serving-cert\") pod \"route-controller-manager-6d679bc5f5-b6c7c\" (UID: \"97e2decb-97dc-49b8-a506-e9009b1e4af1\") " pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.592272 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98cth\" (UniqueName: \"kubernetes.io/projected/97e2decb-97dc-49b8-a506-e9009b1e4af1-kube-api-access-98cth\") pod \"route-controller-manager-6d679bc5f5-b6c7c\" (UID: \"97e2decb-97dc-49b8-a506-e9009b1e4af1\") " pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.593533 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97e2decb-97dc-49b8-a506-e9009b1e4af1-config\") pod \"route-controller-manager-6d679bc5f5-b6c7c\" (UID: \"97e2decb-97dc-49b8-a506-e9009b1e4af1\") " pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.593562 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/97e2decb-97dc-49b8-a506-e9009b1e4af1-client-ca\") pod \"route-controller-manager-6d679bc5f5-b6c7c\" (UID: \"97e2decb-97dc-49b8-a506-e9009b1e4af1\") " pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.601369 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97e2decb-97dc-49b8-a506-e9009b1e4af1-serving-cert\") pod \"route-controller-manager-6d679bc5f5-b6c7c\" (UID: \"97e2decb-97dc-49b8-a506-e9009b1e4af1\") " pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.615034 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98cth\" (UniqueName: \"kubernetes.io/projected/97e2decb-97dc-49b8-a506-e9009b1e4af1-kube-api-access-98cth\") pod \"route-controller-manager-6d679bc5f5-b6c7c\" (UID: \"97e2decb-97dc-49b8-a506-e9009b1e4af1\") " pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:25 crc kubenswrapper[4806]: I1127 10:25:25.630326 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:26 crc kubenswrapper[4806]: I1127 10:25:26.094859 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c"] Nov 27 10:25:26 crc kubenswrapper[4806]: I1127 10:25:26.130906 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cb0392a-70a2-48c0-a69e-4421dbefe86e" path="/var/lib/kubelet/pods/8cb0392a-70a2-48c0-a69e-4421dbefe86e/volumes" Nov 27 10:25:27 crc kubenswrapper[4806]: I1127 10:25:27.001249 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" event={"ID":"97e2decb-97dc-49b8-a506-e9009b1e4af1","Type":"ContainerStarted","Data":"051b0ac0fe44582712829e29bb4a2b9539c8ae79e7aabfe2e890069d73de0ba9"} Nov 27 10:25:27 crc kubenswrapper[4806]: I1127 10:25:27.001770 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" event={"ID":"97e2decb-97dc-49b8-a506-e9009b1e4af1","Type":"ContainerStarted","Data":"5d10e805b53ab4e5fe6134dddf8039bdf7eec96290d131706182cd2070529f24"} Nov 27 10:25:27 crc kubenswrapper[4806]: I1127 10:25:27.031762 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" podStartSLOduration=6.031734601 podStartE2EDuration="6.031734601s" podCreationTimestamp="2025-11-27 10:25:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:25:27.029304412 +0000 UTC m=+231.615895176" watchObservedRunningTime="2025-11-27 10:25:27.031734601 +0000 UTC m=+231.618325365" Nov 27 10:25:28 crc kubenswrapper[4806]: I1127 10:25:28.005722 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:28 crc kubenswrapper[4806]: I1127 10:25:28.011621 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6d679bc5f5-b6c7c" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.183438 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wfw97"] Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.185784 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wfw97" podUID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" containerName="registry-server" containerID="cri-o://21415667b92d8ac1cbcc776bce593aabfe3e50bd2f18f6b58cea510fdc28713b" gracePeriod=30 Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.198963 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jprw7"] Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.199283 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jprw7" podUID="a3bc679e-e16d-47ce-9015-c110b1604ac1" containerName="registry-server" containerID="cri-o://f5ff479bfdd746688aa212d0a72af52cbc5e3451351b9e28b812d6229ade8015" gracePeriod=30 Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.210071 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7t7px"] Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.210458 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" podUID="7ce6a8cd-9071-4c23-a36f-9af525e7f04e" containerName="marketplace-operator" containerID="cri-o://aa8ff45313b16d33e68b2e1c4a5aebbeaa37646cfb844b7cdd41b1e30d68e712" gracePeriod=30 Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.220388 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp6rt"] Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.220705 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gp6rt" podUID="3209cb7e-221c-4c6a-9864-8101e8171078" containerName="registry-server" containerID="cri-o://ab61192ff5176ca7c0370bdd732ef6e16744dd6fc5f396444c78efe04c32275a" gracePeriod=30 Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.239842 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xcj48"] Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.240800 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.254766 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn9nj\" (UniqueName: \"kubernetes.io/projected/4e80f08f-5327-4a27-a182-f26bf02f2f17-kube-api-access-pn9nj\") pod \"marketplace-operator-79b997595-xcj48\" (UID: \"4e80f08f-5327-4a27-a182-f26bf02f2f17\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.254831 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4e80f08f-5327-4a27-a182-f26bf02f2f17-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xcj48\" (UID: \"4e80f08f-5327-4a27-a182-f26bf02f2f17\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.254920 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e80f08f-5327-4a27-a182-f26bf02f2f17-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xcj48\" (UID: \"4e80f08f-5327-4a27-a182-f26bf02f2f17\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.255315 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nmrrm"] Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.255616 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nmrrm" podUID="293a1c8c-8131-462c-9dc4-4a8c516478c5" containerName="registry-server" containerID="cri-o://fd8d33d4b580fcaab375582ed651c86fc547e475440aca5e7c5716bcbdd71bcb" gracePeriod=30 Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.255973 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xcj48"] Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.355937 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn9nj\" (UniqueName: \"kubernetes.io/projected/4e80f08f-5327-4a27-a182-f26bf02f2f17-kube-api-access-pn9nj\") pod \"marketplace-operator-79b997595-xcj48\" (UID: \"4e80f08f-5327-4a27-a182-f26bf02f2f17\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.355992 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4e80f08f-5327-4a27-a182-f26bf02f2f17-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xcj48\" (UID: \"4e80f08f-5327-4a27-a182-f26bf02f2f17\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.356039 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e80f08f-5327-4a27-a182-f26bf02f2f17-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xcj48\" (UID: \"4e80f08f-5327-4a27-a182-f26bf02f2f17\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.360020 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e80f08f-5327-4a27-a182-f26bf02f2f17-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xcj48\" (UID: \"4e80f08f-5327-4a27-a182-f26bf02f2f17\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.378052 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4e80f08f-5327-4a27-a182-f26bf02f2f17-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xcj48\" (UID: \"4e80f08f-5327-4a27-a182-f26bf02f2f17\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.378187 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn9nj\" (UniqueName: \"kubernetes.io/projected/4e80f08f-5327-4a27-a182-f26bf02f2f17-kube-api-access-pn9nj\") pod \"marketplace-operator-79b997595-xcj48\" (UID: \"4e80f08f-5327-4a27-a182-f26bf02f2f17\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.622766 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.704468 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.863540 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf426\" (UniqueName: \"kubernetes.io/projected/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-kube-api-access-tf426\") pod \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\" (UID: \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\") " Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.863591 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-utilities\") pod \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\" (UID: \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\") " Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.863672 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-catalog-content\") pod \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\" (UID: \"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2\") " Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.865147 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-utilities" (OuterVolumeSpecName: "utilities") pod "ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" (UID: "ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.883203 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-kube-api-access-tf426" (OuterVolumeSpecName: "kube-api-access-tf426") pod "ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" (UID: "ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2"). InnerVolumeSpecName "kube-api-access-tf426". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.944034 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" (UID: "ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.966386 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.966433 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf426\" (UniqueName: \"kubernetes.io/projected/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-kube-api-access-tf426\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:36 crc kubenswrapper[4806]: I1127 10:25:36.966451 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.020562 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:25:37 crc kubenswrapper[4806]: E1127 10:25:37.032126 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f5ff479bfdd746688aa212d0a72af52cbc5e3451351b9e28b812d6229ade8015 is running failed: container process not found" containerID="f5ff479bfdd746688aa212d0a72af52cbc5e3451351b9e28b812d6229ade8015" cmd=["grpc_health_probe","-addr=:50051"] Nov 27 10:25:37 crc kubenswrapper[4806]: E1127 10:25:37.033048 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f5ff479bfdd746688aa212d0a72af52cbc5e3451351b9e28b812d6229ade8015 is running failed: container process not found" containerID="f5ff479bfdd746688aa212d0a72af52cbc5e3451351b9e28b812d6229ade8015" cmd=["grpc_health_probe","-addr=:50051"] Nov 27 10:25:37 crc kubenswrapper[4806]: E1127 10:25:37.033994 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f5ff479bfdd746688aa212d0a72af52cbc5e3451351b9e28b812d6229ade8015 is running failed: container process not found" containerID="f5ff479bfdd746688aa212d0a72af52cbc5e3451351b9e28b812d6229ade8015" cmd=["grpc_health_probe","-addr=:50051"] Nov 27 10:25:37 crc kubenswrapper[4806]: E1127 10:25:37.034036 4806 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f5ff479bfdd746688aa212d0a72af52cbc5e3451351b9e28b812d6229ade8015 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-jprw7" podUID="a3bc679e-e16d-47ce-9015-c110b1604ac1" containerName="registry-server" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.128035 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.128837 4806 generic.go:334] "Generic (PLEG): container finished" podID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" containerID="21415667b92d8ac1cbcc776bce593aabfe3e50bd2f18f6b58cea510fdc28713b" exitCode=0 Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.128920 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfw97" event={"ID":"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2","Type":"ContainerDied","Data":"21415667b92d8ac1cbcc776bce593aabfe3e50bd2f18f6b58cea510fdc28713b"} Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.128949 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfw97" event={"ID":"ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2","Type":"ContainerDied","Data":"e10cafd34ba5d265ee51dcdc7936eb1c1779ce58da22bde0c85d96551d80191b"} Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.128979 4806 scope.go:117] "RemoveContainer" containerID="21415667b92d8ac1cbcc776bce593aabfe3e50bd2f18f6b58cea510fdc28713b" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.129120 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfw97" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.155142 4806 generic.go:334] "Generic (PLEG): container finished" podID="293a1c8c-8131-462c-9dc4-4a8c516478c5" containerID="fd8d33d4b580fcaab375582ed651c86fc547e475440aca5e7c5716bcbdd71bcb" exitCode=0 Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.155384 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nmrrm" event={"ID":"293a1c8c-8131-462c-9dc4-4a8c516478c5","Type":"ContainerDied","Data":"fd8d33d4b580fcaab375582ed651c86fc547e475440aca5e7c5716bcbdd71bcb"} Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.158493 4806 generic.go:334] "Generic (PLEG): container finished" podID="7ce6a8cd-9071-4c23-a36f-9af525e7f04e" containerID="aa8ff45313b16d33e68b2e1c4a5aebbeaa37646cfb844b7cdd41b1e30d68e712" exitCode=0 Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.158596 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" event={"ID":"7ce6a8cd-9071-4c23-a36f-9af525e7f04e","Type":"ContainerDied","Data":"aa8ff45313b16d33e68b2e1c4a5aebbeaa37646cfb844b7cdd41b1e30d68e712"} Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.158629 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" event={"ID":"7ce6a8cd-9071-4c23-a36f-9af525e7f04e","Type":"ContainerDied","Data":"ae3e40924632603c92e2c352997042f2f40c55caaf58bc813a613c6130eb2a7e"} Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.158801 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7t7px" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.164209 4806 generic.go:334] "Generic (PLEG): container finished" podID="3209cb7e-221c-4c6a-9864-8101e8171078" containerID="ab61192ff5176ca7c0370bdd732ef6e16744dd6fc5f396444c78efe04c32275a" exitCode=0 Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.164304 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp6rt" event={"ID":"3209cb7e-221c-4c6a-9864-8101e8171078","Type":"ContainerDied","Data":"ab61192ff5176ca7c0370bdd732ef6e16744dd6fc5f396444c78efe04c32275a"} Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.164336 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp6rt" event={"ID":"3209cb7e-221c-4c6a-9864-8101e8171078","Type":"ContainerDied","Data":"e266031e45ae8fdebdbafdf907c955d8a6fed6f604c3aab0e667bcbe7db602ff"} Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.164417 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp6rt" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.167988 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-marketplace-trusted-ca\") pod \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\" (UID: \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\") " Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.168079 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-marketplace-operator-metrics\") pod \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\" (UID: \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\") " Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.168189 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2b8p\" (UniqueName: \"kubernetes.io/projected/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-kube-api-access-d2b8p\") pod \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\" (UID: \"7ce6a8cd-9071-4c23-a36f-9af525e7f04e\") " Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.171743 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-kube-api-access-d2b8p" (OuterVolumeSpecName: "kube-api-access-d2b8p") pod "7ce6a8cd-9071-4c23-a36f-9af525e7f04e" (UID: "7ce6a8cd-9071-4c23-a36f-9af525e7f04e"). InnerVolumeSpecName "kube-api-access-d2b8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.172165 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "7ce6a8cd-9071-4c23-a36f-9af525e7f04e" (UID: "7ce6a8cd-9071-4c23-a36f-9af525e7f04e"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.175736 4806 scope.go:117] "RemoveContainer" containerID="868dfbdf3ed375064d08ec9331cd332b20c5938c07094c0e32e931f9bf5f1646" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.186154 4806 generic.go:334] "Generic (PLEG): container finished" podID="a3bc679e-e16d-47ce-9015-c110b1604ac1" containerID="f5ff479bfdd746688aa212d0a72af52cbc5e3451351b9e28b812d6229ade8015" exitCode=0 Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.186209 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jprw7" event={"ID":"a3bc679e-e16d-47ce-9015-c110b1604ac1","Type":"ContainerDied","Data":"f5ff479bfdd746688aa212d0a72af52cbc5e3451351b9e28b812d6229ade8015"} Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.191658 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "7ce6a8cd-9071-4c23-a36f-9af525e7f04e" (UID: "7ce6a8cd-9071-4c23-a36f-9af525e7f04e"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.225224 4806 scope.go:117] "RemoveContainer" containerID="91e31ee3fcd48668d118ff226542405e05bc0e7a2ba717204372e4d83edd9a0b" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.229341 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wfw97"] Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.241911 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wfw97"] Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.275562 4806 scope.go:117] "RemoveContainer" containerID="21415667b92d8ac1cbcc776bce593aabfe3e50bd2f18f6b58cea510fdc28713b" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.280214 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3209cb7e-221c-4c6a-9864-8101e8171078-catalog-content\") pod \"3209cb7e-221c-4c6a-9864-8101e8171078\" (UID: \"3209cb7e-221c-4c6a-9864-8101e8171078\") " Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.280601 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3209cb7e-221c-4c6a-9864-8101e8171078-utilities\") pod \"3209cb7e-221c-4c6a-9864-8101e8171078\" (UID: \"3209cb7e-221c-4c6a-9864-8101e8171078\") " Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.280822 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5qgh\" (UniqueName: \"kubernetes.io/projected/3209cb7e-221c-4c6a-9864-8101e8171078-kube-api-access-w5qgh\") pod \"3209cb7e-221c-4c6a-9864-8101e8171078\" (UID: \"3209cb7e-221c-4c6a-9864-8101e8171078\") " Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.281605 4806 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.281623 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2b8p\" (UniqueName: \"kubernetes.io/projected/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-kube-api-access-d2b8p\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.281634 4806 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ce6a8cd-9071-4c23-a36f-9af525e7f04e-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.283180 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3209cb7e-221c-4c6a-9864-8101e8171078-utilities" (OuterVolumeSpecName: "utilities") pod "3209cb7e-221c-4c6a-9864-8101e8171078" (UID: "3209cb7e-221c-4c6a-9864-8101e8171078"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:37 crc kubenswrapper[4806]: E1127 10:25:37.290074 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21415667b92d8ac1cbcc776bce593aabfe3e50bd2f18f6b58cea510fdc28713b\": container with ID starting with 21415667b92d8ac1cbcc776bce593aabfe3e50bd2f18f6b58cea510fdc28713b not found: ID does not exist" containerID="21415667b92d8ac1cbcc776bce593aabfe3e50bd2f18f6b58cea510fdc28713b" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.290112 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21415667b92d8ac1cbcc776bce593aabfe3e50bd2f18f6b58cea510fdc28713b"} err="failed to get container status \"21415667b92d8ac1cbcc776bce593aabfe3e50bd2f18f6b58cea510fdc28713b\": rpc error: code = NotFound desc = could not find container \"21415667b92d8ac1cbcc776bce593aabfe3e50bd2f18f6b58cea510fdc28713b\": container with ID starting with 21415667b92d8ac1cbcc776bce593aabfe3e50bd2f18f6b58cea510fdc28713b not found: ID does not exist" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.290141 4806 scope.go:117] "RemoveContainer" containerID="868dfbdf3ed375064d08ec9331cd332b20c5938c07094c0e32e931f9bf5f1646" Nov 27 10:25:37 crc kubenswrapper[4806]: E1127 10:25:37.290702 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"868dfbdf3ed375064d08ec9331cd332b20c5938c07094c0e32e931f9bf5f1646\": container with ID starting with 868dfbdf3ed375064d08ec9331cd332b20c5938c07094c0e32e931f9bf5f1646 not found: ID does not exist" containerID="868dfbdf3ed375064d08ec9331cd332b20c5938c07094c0e32e931f9bf5f1646" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.290758 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"868dfbdf3ed375064d08ec9331cd332b20c5938c07094c0e32e931f9bf5f1646"} err="failed to get container status \"868dfbdf3ed375064d08ec9331cd332b20c5938c07094c0e32e931f9bf5f1646\": rpc error: code = NotFound desc = could not find container \"868dfbdf3ed375064d08ec9331cd332b20c5938c07094c0e32e931f9bf5f1646\": container with ID starting with 868dfbdf3ed375064d08ec9331cd332b20c5938c07094c0e32e931f9bf5f1646 not found: ID does not exist" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.290777 4806 scope.go:117] "RemoveContainer" containerID="91e31ee3fcd48668d118ff226542405e05bc0e7a2ba717204372e4d83edd9a0b" Nov 27 10:25:37 crc kubenswrapper[4806]: E1127 10:25:37.291086 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91e31ee3fcd48668d118ff226542405e05bc0e7a2ba717204372e4d83edd9a0b\": container with ID starting with 91e31ee3fcd48668d118ff226542405e05bc0e7a2ba717204372e4d83edd9a0b not found: ID does not exist" containerID="91e31ee3fcd48668d118ff226542405e05bc0e7a2ba717204372e4d83edd9a0b" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.291104 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e31ee3fcd48668d118ff226542405e05bc0e7a2ba717204372e4d83edd9a0b"} err="failed to get container status \"91e31ee3fcd48668d118ff226542405e05bc0e7a2ba717204372e4d83edd9a0b\": rpc error: code = NotFound desc = could not find container \"91e31ee3fcd48668d118ff226542405e05bc0e7a2ba717204372e4d83edd9a0b\": container with ID starting with 91e31ee3fcd48668d118ff226542405e05bc0e7a2ba717204372e4d83edd9a0b not found: ID does not exist" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.291118 4806 scope.go:117] "RemoveContainer" containerID="aa8ff45313b16d33e68b2e1c4a5aebbeaa37646cfb844b7cdd41b1e30d68e712" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.294642 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3209cb7e-221c-4c6a-9864-8101e8171078-kube-api-access-w5qgh" (OuterVolumeSpecName: "kube-api-access-w5qgh") pod "3209cb7e-221c-4c6a-9864-8101e8171078" (UID: "3209cb7e-221c-4c6a-9864-8101e8171078"). InnerVolumeSpecName "kube-api-access-w5qgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.318375 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3209cb7e-221c-4c6a-9864-8101e8171078-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3209cb7e-221c-4c6a-9864-8101e8171078" (UID: "3209cb7e-221c-4c6a-9864-8101e8171078"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.343631 4806 scope.go:117] "RemoveContainer" containerID="aa8ff45313b16d33e68b2e1c4a5aebbeaa37646cfb844b7cdd41b1e30d68e712" Nov 27 10:25:37 crc kubenswrapper[4806]: E1127 10:25:37.344064 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa8ff45313b16d33e68b2e1c4a5aebbeaa37646cfb844b7cdd41b1e30d68e712\": container with ID starting with aa8ff45313b16d33e68b2e1c4a5aebbeaa37646cfb844b7cdd41b1e30d68e712 not found: ID does not exist" containerID="aa8ff45313b16d33e68b2e1c4a5aebbeaa37646cfb844b7cdd41b1e30d68e712" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.344101 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa8ff45313b16d33e68b2e1c4a5aebbeaa37646cfb844b7cdd41b1e30d68e712"} err="failed to get container status \"aa8ff45313b16d33e68b2e1c4a5aebbeaa37646cfb844b7cdd41b1e30d68e712\": rpc error: code = NotFound desc = could not find container \"aa8ff45313b16d33e68b2e1c4a5aebbeaa37646cfb844b7cdd41b1e30d68e712\": container with ID starting with aa8ff45313b16d33e68b2e1c4a5aebbeaa37646cfb844b7cdd41b1e30d68e712 not found: ID does not exist" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.344134 4806 scope.go:117] "RemoveContainer" containerID="ab61192ff5176ca7c0370bdd732ef6e16744dd6fc5f396444c78efe04c32275a" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.383836 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3209cb7e-221c-4c6a-9864-8101e8171078-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.383869 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5qgh\" (UniqueName: \"kubernetes.io/projected/3209cb7e-221c-4c6a-9864-8101e8171078-kube-api-access-w5qgh\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.383882 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3209cb7e-221c-4c6a-9864-8101e8171078-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.425168 4806 scope.go:117] "RemoveContainer" containerID="ddcb136b375f253c58a463268b9106aceb9a47d857a54e3c3a790116ee1018e0" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.452369 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xcj48"] Nov 27 10:25:37 crc kubenswrapper[4806]: W1127 10:25:37.492059 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e80f08f_5327_4a27_a182_f26bf02f2f17.slice/crio-7357064bb062d63e7ffccfd91620722d7e914db564b710808cdbd01e2c47a609 WatchSource:0}: Error finding container 7357064bb062d63e7ffccfd91620722d7e914db564b710808cdbd01e2c47a609: Status 404 returned error can't find the container with id 7357064bb062d63e7ffccfd91620722d7e914db564b710808cdbd01e2c47a609 Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.492192 4806 scope.go:117] "RemoveContainer" containerID="485f96fbbca01e2926f50422a4174a0874318ee9463dbbe63675bc94bf86d2e9" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.525042 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7t7px"] Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.532005 4806 scope.go:117] "RemoveContainer" containerID="ab61192ff5176ca7c0370bdd732ef6e16744dd6fc5f396444c78efe04c32275a" Nov 27 10:25:37 crc kubenswrapper[4806]: E1127 10:25:37.534037 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab61192ff5176ca7c0370bdd732ef6e16744dd6fc5f396444c78efe04c32275a\": container with ID starting with ab61192ff5176ca7c0370bdd732ef6e16744dd6fc5f396444c78efe04c32275a not found: ID does not exist" containerID="ab61192ff5176ca7c0370bdd732ef6e16744dd6fc5f396444c78efe04c32275a" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.534089 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab61192ff5176ca7c0370bdd732ef6e16744dd6fc5f396444c78efe04c32275a"} err="failed to get container status \"ab61192ff5176ca7c0370bdd732ef6e16744dd6fc5f396444c78efe04c32275a\": rpc error: code = NotFound desc = could not find container \"ab61192ff5176ca7c0370bdd732ef6e16744dd6fc5f396444c78efe04c32275a\": container with ID starting with ab61192ff5176ca7c0370bdd732ef6e16744dd6fc5f396444c78efe04c32275a not found: ID does not exist" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.534121 4806 scope.go:117] "RemoveContainer" containerID="ddcb136b375f253c58a463268b9106aceb9a47d857a54e3c3a790116ee1018e0" Nov 27 10:25:37 crc kubenswrapper[4806]: E1127 10:25:37.534395 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddcb136b375f253c58a463268b9106aceb9a47d857a54e3c3a790116ee1018e0\": container with ID starting with ddcb136b375f253c58a463268b9106aceb9a47d857a54e3c3a790116ee1018e0 not found: ID does not exist" containerID="ddcb136b375f253c58a463268b9106aceb9a47d857a54e3c3a790116ee1018e0" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.534412 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddcb136b375f253c58a463268b9106aceb9a47d857a54e3c3a790116ee1018e0"} err="failed to get container status \"ddcb136b375f253c58a463268b9106aceb9a47d857a54e3c3a790116ee1018e0\": rpc error: code = NotFound desc = could not find container \"ddcb136b375f253c58a463268b9106aceb9a47d857a54e3c3a790116ee1018e0\": container with ID starting with ddcb136b375f253c58a463268b9106aceb9a47d857a54e3c3a790116ee1018e0 not found: ID does not exist" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.534424 4806 scope.go:117] "RemoveContainer" containerID="485f96fbbca01e2926f50422a4174a0874318ee9463dbbe63675bc94bf86d2e9" Nov 27 10:25:37 crc kubenswrapper[4806]: E1127 10:25:37.536745 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"485f96fbbca01e2926f50422a4174a0874318ee9463dbbe63675bc94bf86d2e9\": container with ID starting with 485f96fbbca01e2926f50422a4174a0874318ee9463dbbe63675bc94bf86d2e9 not found: ID does not exist" containerID="485f96fbbca01e2926f50422a4174a0874318ee9463dbbe63675bc94bf86d2e9" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.536903 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"485f96fbbca01e2926f50422a4174a0874318ee9463dbbe63675bc94bf86d2e9"} err="failed to get container status \"485f96fbbca01e2926f50422a4174a0874318ee9463dbbe63675bc94bf86d2e9\": rpc error: code = NotFound desc = could not find container \"485f96fbbca01e2926f50422a4174a0874318ee9463dbbe63675bc94bf86d2e9\": container with ID starting with 485f96fbbca01e2926f50422a4174a0874318ee9463dbbe63675bc94bf86d2e9 not found: ID does not exist" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.546920 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7t7px"] Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.557516 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp6rt"] Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.557819 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp6rt"] Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.672509 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.680051 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.803065 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3bc679e-e16d-47ce-9015-c110b1604ac1-catalog-content\") pod \"a3bc679e-e16d-47ce-9015-c110b1604ac1\" (UID: \"a3bc679e-e16d-47ce-9015-c110b1604ac1\") " Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.803130 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbvv8\" (UniqueName: \"kubernetes.io/projected/293a1c8c-8131-462c-9dc4-4a8c516478c5-kube-api-access-rbvv8\") pod \"293a1c8c-8131-462c-9dc4-4a8c516478c5\" (UID: \"293a1c8c-8131-462c-9dc4-4a8c516478c5\") " Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.803170 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3bc679e-e16d-47ce-9015-c110b1604ac1-utilities\") pod \"a3bc679e-e16d-47ce-9015-c110b1604ac1\" (UID: \"a3bc679e-e16d-47ce-9015-c110b1604ac1\") " Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.803197 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/293a1c8c-8131-462c-9dc4-4a8c516478c5-utilities\") pod \"293a1c8c-8131-462c-9dc4-4a8c516478c5\" (UID: \"293a1c8c-8131-462c-9dc4-4a8c516478c5\") " Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.803266 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw5vm\" (UniqueName: \"kubernetes.io/projected/a3bc679e-e16d-47ce-9015-c110b1604ac1-kube-api-access-zw5vm\") pod \"a3bc679e-e16d-47ce-9015-c110b1604ac1\" (UID: \"a3bc679e-e16d-47ce-9015-c110b1604ac1\") " Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.803308 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/293a1c8c-8131-462c-9dc4-4a8c516478c5-catalog-content\") pod \"293a1c8c-8131-462c-9dc4-4a8c516478c5\" (UID: \"293a1c8c-8131-462c-9dc4-4a8c516478c5\") " Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.804376 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/293a1c8c-8131-462c-9dc4-4a8c516478c5-utilities" (OuterVolumeSpecName: "utilities") pod "293a1c8c-8131-462c-9dc4-4a8c516478c5" (UID: "293a1c8c-8131-462c-9dc4-4a8c516478c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.805117 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3bc679e-e16d-47ce-9015-c110b1604ac1-utilities" (OuterVolumeSpecName: "utilities") pod "a3bc679e-e16d-47ce-9015-c110b1604ac1" (UID: "a3bc679e-e16d-47ce-9015-c110b1604ac1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.809120 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/293a1c8c-8131-462c-9dc4-4a8c516478c5-kube-api-access-rbvv8" (OuterVolumeSpecName: "kube-api-access-rbvv8") pod "293a1c8c-8131-462c-9dc4-4a8c516478c5" (UID: "293a1c8c-8131-462c-9dc4-4a8c516478c5"). InnerVolumeSpecName "kube-api-access-rbvv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.809204 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3bc679e-e16d-47ce-9015-c110b1604ac1-kube-api-access-zw5vm" (OuterVolumeSpecName: "kube-api-access-zw5vm") pod "a3bc679e-e16d-47ce-9015-c110b1604ac1" (UID: "a3bc679e-e16d-47ce-9015-c110b1604ac1"). InnerVolumeSpecName "kube-api-access-zw5vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.863836 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3bc679e-e16d-47ce-9015-c110b1604ac1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3bc679e-e16d-47ce-9015-c110b1604ac1" (UID: "a3bc679e-e16d-47ce-9015-c110b1604ac1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.905371 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbvv8\" (UniqueName: \"kubernetes.io/projected/293a1c8c-8131-462c-9dc4-4a8c516478c5-kube-api-access-rbvv8\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.905411 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3bc679e-e16d-47ce-9015-c110b1604ac1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.905423 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3bc679e-e16d-47ce-9015-c110b1604ac1-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.905433 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/293a1c8c-8131-462c-9dc4-4a8c516478c5-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.905443 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw5vm\" (UniqueName: \"kubernetes.io/projected/a3bc679e-e16d-47ce-9015-c110b1604ac1-kube-api-access-zw5vm\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:37 crc kubenswrapper[4806]: I1127 10:25:37.906173 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/293a1c8c-8131-462c-9dc4-4a8c516478c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "293a1c8c-8131-462c-9dc4-4a8c516478c5" (UID: "293a1c8c-8131-462c-9dc4-4a8c516478c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.006494 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/293a1c8c-8131-462c-9dc4-4a8c516478c5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.126027 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3209cb7e-221c-4c6a-9864-8101e8171078" path="/var/lib/kubelet/pods/3209cb7e-221c-4c6a-9864-8101e8171078/volumes" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.127335 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ce6a8cd-9071-4c23-a36f-9af525e7f04e" path="/var/lib/kubelet/pods/7ce6a8cd-9071-4c23-a36f-9af525e7f04e/volumes" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.128158 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" path="/var/lib/kubelet/pods/ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2/volumes" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.193396 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" event={"ID":"4e80f08f-5327-4a27-a182-f26bf02f2f17","Type":"ContainerStarted","Data":"b47c05a2b3b2172af7a5821fb1855fc8c95be893e0dedd3343996c8fa756f377"} Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.193460 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" event={"ID":"4e80f08f-5327-4a27-a182-f26bf02f2f17","Type":"ContainerStarted","Data":"7357064bb062d63e7ffccfd91620722d7e914db564b710808cdbd01e2c47a609"} Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.194138 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.197730 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jprw7" event={"ID":"a3bc679e-e16d-47ce-9015-c110b1604ac1","Type":"ContainerDied","Data":"8d5ee725a6b61939fce7eb307e1e96757a3d4f8c5cff037c49250d44d8ac6fb8"} Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.197784 4806 scope.go:117] "RemoveContainer" containerID="f5ff479bfdd746688aa212d0a72af52cbc5e3451351b9e28b812d6229ade8015" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.197887 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jprw7" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.204003 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nmrrm" event={"ID":"293a1c8c-8131-462c-9dc4-4a8c516478c5","Type":"ContainerDied","Data":"00034d52ab8eb1c2f52a74cc1ac329e234ca8cf520ec193d6dbbff536036f3f7"} Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.204020 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nmrrm" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.207615 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.217933 4806 scope.go:117] "RemoveContainer" containerID="2120d4c87d3cd8a3b7f1b50c681960d0f4f20ee99282202ad4bff85e20c411db" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.221676 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xcj48" podStartSLOduration=2.221658249 podStartE2EDuration="2.221658249s" podCreationTimestamp="2025-11-27 10:25:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:25:38.218647292 +0000 UTC m=+242.805238046" watchObservedRunningTime="2025-11-27 10:25:38.221658249 +0000 UTC m=+242.808249013" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.239816 4806 scope.go:117] "RemoveContainer" containerID="9802e49e13468f6bc17d0afb1d7c68adc57f35aa536d4437bdf25bec0929fadf" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.257668 4806 scope.go:117] "RemoveContainer" containerID="fd8d33d4b580fcaab375582ed651c86fc547e475440aca5e7c5716bcbdd71bcb" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.261737 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jprw7"] Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.264155 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jprw7"] Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.279913 4806 scope.go:117] "RemoveContainer" containerID="db6a9629a43f497636805a0771797b16ca4fca530a30ce38976b5bff602fe326" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.283153 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nmrrm"] Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.290648 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nmrrm"] Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.295710 4806 scope.go:117] "RemoveContainer" containerID="f8db139004fbced6a3df6579096a70ac39008713c096b920e2cdf3aefcf44f49" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400460 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j54w9"] Nov 27 10:25:38 crc kubenswrapper[4806]: E1127 10:25:38.400672 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce6a8cd-9071-4c23-a36f-9af525e7f04e" containerName="marketplace-operator" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400685 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce6a8cd-9071-4c23-a36f-9af525e7f04e" containerName="marketplace-operator" Nov 27 10:25:38 crc kubenswrapper[4806]: E1127 10:25:38.400696 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3bc679e-e16d-47ce-9015-c110b1604ac1" containerName="extract-utilities" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400702 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3bc679e-e16d-47ce-9015-c110b1604ac1" containerName="extract-utilities" Nov 27 10:25:38 crc kubenswrapper[4806]: E1127 10:25:38.400715 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3bc679e-e16d-47ce-9015-c110b1604ac1" containerName="registry-server" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400720 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3bc679e-e16d-47ce-9015-c110b1604ac1" containerName="registry-server" Nov 27 10:25:38 crc kubenswrapper[4806]: E1127 10:25:38.400727 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" containerName="registry-server" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400732 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" containerName="registry-server" Nov 27 10:25:38 crc kubenswrapper[4806]: E1127 10:25:38.400741 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="293a1c8c-8131-462c-9dc4-4a8c516478c5" containerName="registry-server" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400750 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="293a1c8c-8131-462c-9dc4-4a8c516478c5" containerName="registry-server" Nov 27 10:25:38 crc kubenswrapper[4806]: E1127 10:25:38.400759 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3209cb7e-221c-4c6a-9864-8101e8171078" containerName="extract-content" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400765 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3209cb7e-221c-4c6a-9864-8101e8171078" containerName="extract-content" Nov 27 10:25:38 crc kubenswrapper[4806]: E1127 10:25:38.400774 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3209cb7e-221c-4c6a-9864-8101e8171078" containerName="registry-server" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400780 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3209cb7e-221c-4c6a-9864-8101e8171078" containerName="registry-server" Nov 27 10:25:38 crc kubenswrapper[4806]: E1127 10:25:38.400790 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" containerName="extract-utilities" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400795 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" containerName="extract-utilities" Nov 27 10:25:38 crc kubenswrapper[4806]: E1127 10:25:38.400803 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3209cb7e-221c-4c6a-9864-8101e8171078" containerName="extract-utilities" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400810 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3209cb7e-221c-4c6a-9864-8101e8171078" containerName="extract-utilities" Nov 27 10:25:38 crc kubenswrapper[4806]: E1127 10:25:38.400816 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="293a1c8c-8131-462c-9dc4-4a8c516478c5" containerName="extract-content" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400821 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="293a1c8c-8131-462c-9dc4-4a8c516478c5" containerName="extract-content" Nov 27 10:25:38 crc kubenswrapper[4806]: E1127 10:25:38.400829 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="293a1c8c-8131-462c-9dc4-4a8c516478c5" containerName="extract-utilities" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400835 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="293a1c8c-8131-462c-9dc4-4a8c516478c5" containerName="extract-utilities" Nov 27 10:25:38 crc kubenswrapper[4806]: E1127 10:25:38.400845 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3bc679e-e16d-47ce-9015-c110b1604ac1" containerName="extract-content" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400852 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3bc679e-e16d-47ce-9015-c110b1604ac1" containerName="extract-content" Nov 27 10:25:38 crc kubenswrapper[4806]: E1127 10:25:38.400862 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" containerName="extract-content" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400868 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" containerName="extract-content" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400958 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce6a8cd-9071-4c23-a36f-9af525e7f04e" containerName="marketplace-operator" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400970 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba6aacd9-e39c-4a2c-bcfc-f5d1c252bdb2" containerName="registry-server" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400982 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3209cb7e-221c-4c6a-9864-8101e8171078" containerName="registry-server" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400988 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3bc679e-e16d-47ce-9015-c110b1604ac1" containerName="registry-server" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.400997 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="293a1c8c-8131-462c-9dc4-4a8c516478c5" containerName="registry-server" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.401789 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.404143 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.414852 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j54w9"] Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.514017 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e9f754d-2097-4526-9123-757c1af2a591-catalog-content\") pod \"redhat-marketplace-j54w9\" (UID: \"5e9f754d-2097-4526-9123-757c1af2a591\") " pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.514641 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e9f754d-2097-4526-9123-757c1af2a591-utilities\") pod \"redhat-marketplace-j54w9\" (UID: \"5e9f754d-2097-4526-9123-757c1af2a591\") " pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.514773 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5rg8\" (UniqueName: \"kubernetes.io/projected/5e9f754d-2097-4526-9123-757c1af2a591-kube-api-access-c5rg8\") pod \"redhat-marketplace-j54w9\" (UID: \"5e9f754d-2097-4526-9123-757c1af2a591\") " pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.605390 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wfxvq"] Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.607202 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.612374 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wfxvq"] Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.613128 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.615671 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e9f754d-2097-4526-9123-757c1af2a591-utilities\") pod \"redhat-marketplace-j54w9\" (UID: \"5e9f754d-2097-4526-9123-757c1af2a591\") " pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.615938 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5rg8\" (UniqueName: \"kubernetes.io/projected/5e9f754d-2097-4526-9123-757c1af2a591-kube-api-access-c5rg8\") pod \"redhat-marketplace-j54w9\" (UID: \"5e9f754d-2097-4526-9123-757c1af2a591\") " pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.616366 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e9f754d-2097-4526-9123-757c1af2a591-catalog-content\") pod \"redhat-marketplace-j54w9\" (UID: \"5e9f754d-2097-4526-9123-757c1af2a591\") " pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.616535 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e9f754d-2097-4526-9123-757c1af2a591-utilities\") pod \"redhat-marketplace-j54w9\" (UID: \"5e9f754d-2097-4526-9123-757c1af2a591\") " pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.616754 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e9f754d-2097-4526-9123-757c1af2a591-catalog-content\") pod \"redhat-marketplace-j54w9\" (UID: \"5e9f754d-2097-4526-9123-757c1af2a591\") " pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.641008 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5rg8\" (UniqueName: \"kubernetes.io/projected/5e9f754d-2097-4526-9123-757c1af2a591-kube-api-access-c5rg8\") pod \"redhat-marketplace-j54w9\" (UID: \"5e9f754d-2097-4526-9123-757c1af2a591\") " pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.718174 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-utilities\") pod \"certified-operators-wfxvq\" (UID: \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\") " pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.718296 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-catalog-content\") pod \"certified-operators-wfxvq\" (UID: \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\") " pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.718321 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88rp5\" (UniqueName: \"kubernetes.io/projected/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-kube-api-access-88rp5\") pod \"certified-operators-wfxvq\" (UID: \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\") " pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.754660 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.819994 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-utilities\") pod \"certified-operators-wfxvq\" (UID: \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\") " pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.820068 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-catalog-content\") pod \"certified-operators-wfxvq\" (UID: \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\") " pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.820088 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88rp5\" (UniqueName: \"kubernetes.io/projected/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-kube-api-access-88rp5\") pod \"certified-operators-wfxvq\" (UID: \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\") " pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.820533 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-utilities\") pod \"certified-operators-wfxvq\" (UID: \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\") " pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.820695 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-catalog-content\") pod \"certified-operators-wfxvq\" (UID: \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\") " pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.844855 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88rp5\" (UniqueName: \"kubernetes.io/projected/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-kube-api-access-88rp5\") pod \"certified-operators-wfxvq\" (UID: \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\") " pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:38 crc kubenswrapper[4806]: I1127 10:25:38.930718 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:39 crc kubenswrapper[4806]: I1127 10:25:39.155459 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j54w9"] Nov 27 10:25:39 crc kubenswrapper[4806]: W1127 10:25:39.161155 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e9f754d_2097_4526_9123_757c1af2a591.slice/crio-c42a2fb9ca77d0c718a338d9581d8c3fc32fc2d2a745440988989bdf50265090 WatchSource:0}: Error finding container c42a2fb9ca77d0c718a338d9581d8c3fc32fc2d2a745440988989bdf50265090: Status 404 returned error can't find the container with id c42a2fb9ca77d0c718a338d9581d8c3fc32fc2d2a745440988989bdf50265090 Nov 27 10:25:39 crc kubenswrapper[4806]: I1127 10:25:39.214871 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j54w9" event={"ID":"5e9f754d-2097-4526-9123-757c1af2a591","Type":"ContainerStarted","Data":"c42a2fb9ca77d0c718a338d9581d8c3fc32fc2d2a745440988989bdf50265090"} Nov 27 10:25:39 crc kubenswrapper[4806]: I1127 10:25:39.349062 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wfxvq"] Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.129897 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="293a1c8c-8131-462c-9dc4-4a8c516478c5" path="/var/lib/kubelet/pods/293a1c8c-8131-462c-9dc4-4a8c516478c5/volumes" Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.131694 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3bc679e-e16d-47ce-9015-c110b1604ac1" path="/var/lib/kubelet/pods/a3bc679e-e16d-47ce-9015-c110b1604ac1/volumes" Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.223561 4806 generic.go:334] "Generic (PLEG): container finished" podID="5e9f754d-2097-4526-9123-757c1af2a591" containerID="36c0a46a5c9632672026675d146c5434cc48fc8f56403abcb91d5945524da62c" exitCode=0 Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.223682 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j54w9" event={"ID":"5e9f754d-2097-4526-9123-757c1af2a591","Type":"ContainerDied","Data":"36c0a46a5c9632672026675d146c5434cc48fc8f56403abcb91d5945524da62c"} Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.225575 4806 generic.go:334] "Generic (PLEG): container finished" podID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" containerID="94e79ff145bb937ccf9506389b53e3a77f89371f64e7c94d1ddf76d5f32220e6" exitCode=0 Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.225667 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfxvq" event={"ID":"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3","Type":"ContainerDied","Data":"94e79ff145bb937ccf9506389b53e3a77f89371f64e7c94d1ddf76d5f32220e6"} Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.225706 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfxvq" event={"ID":"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3","Type":"ContainerStarted","Data":"3076fadda0d099ff25c6341026a33f7f0f377f5020968fb1a91de7fa56201a1b"} Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.801109 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-69nx8"] Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.802119 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.805800 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.821174 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-69nx8"] Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.857527 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eff0fde1-b975-41bc-95a4-080177acf5ef-catalog-content\") pod \"redhat-operators-69nx8\" (UID: \"eff0fde1-b975-41bc-95a4-080177acf5ef\") " pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.857604 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4btd\" (UniqueName: \"kubernetes.io/projected/eff0fde1-b975-41bc-95a4-080177acf5ef-kube-api-access-g4btd\") pod \"redhat-operators-69nx8\" (UID: \"eff0fde1-b975-41bc-95a4-080177acf5ef\") " pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.857695 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eff0fde1-b975-41bc-95a4-080177acf5ef-utilities\") pod \"redhat-operators-69nx8\" (UID: \"eff0fde1-b975-41bc-95a4-080177acf5ef\") " pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.958959 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eff0fde1-b975-41bc-95a4-080177acf5ef-catalog-content\") pod \"redhat-operators-69nx8\" (UID: \"eff0fde1-b975-41bc-95a4-080177acf5ef\") " pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.959020 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4btd\" (UniqueName: \"kubernetes.io/projected/eff0fde1-b975-41bc-95a4-080177acf5ef-kube-api-access-g4btd\") pod \"redhat-operators-69nx8\" (UID: \"eff0fde1-b975-41bc-95a4-080177acf5ef\") " pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.959043 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eff0fde1-b975-41bc-95a4-080177acf5ef-utilities\") pod \"redhat-operators-69nx8\" (UID: \"eff0fde1-b975-41bc-95a4-080177acf5ef\") " pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.959543 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eff0fde1-b975-41bc-95a4-080177acf5ef-utilities\") pod \"redhat-operators-69nx8\" (UID: \"eff0fde1-b975-41bc-95a4-080177acf5ef\") " pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.959756 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eff0fde1-b975-41bc-95a4-080177acf5ef-catalog-content\") pod \"redhat-operators-69nx8\" (UID: \"eff0fde1-b975-41bc-95a4-080177acf5ef\") " pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:40 crc kubenswrapper[4806]: I1127 10:25:40.996523 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4btd\" (UniqueName: \"kubernetes.io/projected/eff0fde1-b975-41bc-95a4-080177acf5ef-kube-api-access-g4btd\") pod \"redhat-operators-69nx8\" (UID: \"eff0fde1-b975-41bc-95a4-080177acf5ef\") " pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.008855 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jh6qx"] Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.010317 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.023550 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.032531 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jh6qx"] Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.065844 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct9jr\" (UniqueName: \"kubernetes.io/projected/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-kube-api-access-ct9jr\") pod \"community-operators-jh6qx\" (UID: \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\") " pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.065924 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-catalog-content\") pod \"community-operators-jh6qx\" (UID: \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\") " pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.065985 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-utilities\") pod \"community-operators-jh6qx\" (UID: \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\") " pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.119578 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.167320 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-catalog-content\") pod \"community-operators-jh6qx\" (UID: \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\") " pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.167389 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-utilities\") pod \"community-operators-jh6qx\" (UID: \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\") " pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.167466 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct9jr\" (UniqueName: \"kubernetes.io/projected/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-kube-api-access-ct9jr\") pod \"community-operators-jh6qx\" (UID: \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\") " pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.168029 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-utilities\") pod \"community-operators-jh6qx\" (UID: \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\") " pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.168032 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-catalog-content\") pod \"community-operators-jh6qx\" (UID: \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\") " pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.187663 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct9jr\" (UniqueName: \"kubernetes.io/projected/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-kube-api-access-ct9jr\") pod \"community-operators-jh6qx\" (UID: \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\") " pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.355550 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.621805 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-69nx8"] Nov 27 10:25:41 crc kubenswrapper[4806]: W1127 10:25:41.663415 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeff0fde1_b975_41bc_95a4_080177acf5ef.slice/crio-0c61f8b8d9903cdd6249a8845944da2317b906b847deae4f7ea61fbdd537d096 WatchSource:0}: Error finding container 0c61f8b8d9903cdd6249a8845944da2317b906b847deae4f7ea61fbdd537d096: Status 404 returned error can't find the container with id 0c61f8b8d9903cdd6249a8845944da2317b906b847deae4f7ea61fbdd537d096 Nov 27 10:25:41 crc kubenswrapper[4806]: I1127 10:25:41.856421 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jh6qx"] Nov 27 10:25:41 crc kubenswrapper[4806]: W1127 10:25:41.889415 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7c21192_8990_4c0d_bd2a_e3ba5c1ba11d.slice/crio-5e04dea91ab4046f41d4659c58eea636d8b65efafa5300dbb7de84ae1a836597 WatchSource:0}: Error finding container 5e04dea91ab4046f41d4659c58eea636d8b65efafa5300dbb7de84ae1a836597: Status 404 returned error can't find the container with id 5e04dea91ab4046f41d4659c58eea636d8b65efafa5300dbb7de84ae1a836597 Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.290322 4806 generic.go:334] "Generic (PLEG): container finished" podID="5e9f754d-2097-4526-9123-757c1af2a591" containerID="9aeb9d56303cecb45fb7951c58331bd02e74ae2bb4b1f80a63e2cbf18b769335" exitCode=0 Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.290745 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j54w9" event={"ID":"5e9f754d-2097-4526-9123-757c1af2a591","Type":"ContainerDied","Data":"9aeb9d56303cecb45fb7951c58331bd02e74ae2bb4b1f80a63e2cbf18b769335"} Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.299863 4806 generic.go:334] "Generic (PLEG): container finished" podID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" containerID="d8481b180e70933acad600bdc4e2e433f91c609578f9d03677cfa267f21936f5" exitCode=0 Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.299943 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfxvq" event={"ID":"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3","Type":"ContainerDied","Data":"d8481b180e70933acad600bdc4e2e433f91c609578f9d03677cfa267f21936f5"} Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.308344 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69nx8" event={"ID":"eff0fde1-b975-41bc-95a4-080177acf5ef","Type":"ContainerDied","Data":"1719a75088e15f90112d266a2040bd970653e1851ae055057d26272be044939a"} Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.308823 4806 generic.go:334] "Generic (PLEG): container finished" podID="eff0fde1-b975-41bc-95a4-080177acf5ef" containerID="1719a75088e15f90112d266a2040bd970653e1851ae055057d26272be044939a" exitCode=0 Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.308913 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69nx8" event={"ID":"eff0fde1-b975-41bc-95a4-080177acf5ef","Type":"ContainerStarted","Data":"0c61f8b8d9903cdd6249a8845944da2317b906b847deae4f7ea61fbdd537d096"} Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.314730 4806 generic.go:334] "Generic (PLEG): container finished" podID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" containerID="c249c10cd2c23aa758eee36e2a9ef2ca465977c37fe93b2d3aa0cf104f924030" exitCode=0 Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.314839 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh6qx" event={"ID":"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d","Type":"ContainerDied","Data":"c249c10cd2c23aa758eee36e2a9ef2ca465977c37fe93b2d3aa0cf104f924030"} Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.314891 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh6qx" event={"ID":"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d","Type":"ContainerStarted","Data":"5e04dea91ab4046f41d4659c58eea636d8b65efafa5300dbb7de84ae1a836597"} Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.656433 4806 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.659846 4806 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.660084 4806 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.660426 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61" gracePeriod=15 Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.660688 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280" gracePeriod=15 Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.660869 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411" gracePeriod=15 Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.660810 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310" gracePeriod=15 Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.660010 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.660758 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12" gracePeriod=15 Nov 27 10:25:42 crc kubenswrapper[4806]: E1127 10:25:42.661749 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.661927 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 27 10:25:42 crc kubenswrapper[4806]: E1127 10:25:42.662149 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.662294 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 27 10:25:42 crc kubenswrapper[4806]: E1127 10:25:42.662509 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.662731 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 27 10:25:42 crc kubenswrapper[4806]: E1127 10:25:42.662878 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.662994 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 27 10:25:42 crc kubenswrapper[4806]: E1127 10:25:42.663107 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.663320 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 10:25:42 crc kubenswrapper[4806]: E1127 10:25:42.663497 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.663634 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 27 10:25:42 crc kubenswrapper[4806]: E1127 10:25:42.663765 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.663880 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.664172 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.664352 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.664621 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.664767 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.665063 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.666836 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.665806 4806 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.693378 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.694788 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.694961 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.703373 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.703478 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.703530 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.703564 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.703737 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.805780 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.805853 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.805882 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.805902 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.805982 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.805987 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.806005 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.806088 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.806103 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.806131 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.806208 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.806027 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.807192 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.807216 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.807343 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: I1127 10:25:42.807341 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:42 crc kubenswrapper[4806]: E1127 10:25:42.970337 4806 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.246:6443: connect: connection refused" event="&Event{ObjectMeta:{certified-operators-wfxvq.187bd62737861c08 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:certified-operators-wfxvq,UID:dcdfc5eb-7543-43b1-bbb5-90734e6f85a3,APIVersion:v1,ResourceVersion:29696,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 664ms (664ms including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 10:25:42.96928564 +0000 UTC m=+247.555876404,LastTimestamp:2025-11-27 10:25:42.96928564 +0000 UTC m=+247.555876404,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.324762 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j54w9" event={"ID":"5e9f754d-2097-4526-9123-757c1af2a591","Type":"ContainerStarted","Data":"9fbf9df6498bd647062eaaba3114d3abc5df3c32eef3a2c65512e6c25608ea5d"} Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.325988 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.328285 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfxvq" event={"ID":"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3","Type":"ContainerStarted","Data":"5eb7db305ecb246b3fd9bafaac53f10afac8a86ea08e3245141892892620d3f3"} Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.329078 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.329674 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.332298 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh6qx" event={"ID":"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d","Type":"ContainerStarted","Data":"f295f1e491c669bef133f65ff7de5819cae5c506e39d7507a6f15d5b0759d42a"} Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.333548 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.333931 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.334335 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.338517 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.340340 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.343867 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411" exitCode=0 Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.343889 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280" exitCode=0 Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.343898 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310" exitCode=0 Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.343908 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12" exitCode=2 Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.344023 4806 scope.go:117] "RemoveContainer" containerID="355b0218dd780513fa87ab6023994297bf58e0868ef60bafb8950ce0e6a09127" Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.349360 4806 generic.go:334] "Generic (PLEG): container finished" podID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" containerID="3ce58a2a37537acf3a2b33025573e96d72e0ce8f4070481d5b5cf22d13f601aa" exitCode=0 Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.349418 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"38a72bf5-a9ca-409a-b52d-b10ebefc316a","Type":"ContainerDied","Data":"3ce58a2a37537acf3a2b33025573e96d72e0ce8f4070481d5b5cf22d13f601aa"} Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.350247 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.350623 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.350945 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:43 crc kubenswrapper[4806]: I1127 10:25:43.351272 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.358361 4806 generic.go:334] "Generic (PLEG): container finished" podID="eff0fde1-b975-41bc-95a4-080177acf5ef" containerID="7c7bce4f59a6d5dd95425a3b12553d81542dc9479491a14c2a1d282faa6a3d31" exitCode=0 Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.358989 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69nx8" event={"ID":"eff0fde1-b975-41bc-95a4-080177acf5ef","Type":"ContainerDied","Data":"7c7bce4f59a6d5dd95425a3b12553d81542dc9479491a14c2a1d282faa6a3d31"} Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.360384 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.360618 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.360950 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.361351 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.361566 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.366124 4806 generic.go:334] "Generic (PLEG): container finished" podID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" containerID="f295f1e491c669bef133f65ff7de5819cae5c506e39d7507a6f15d5b0759d42a" exitCode=0 Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.366269 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh6qx" event={"ID":"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d","Type":"ContainerDied","Data":"f295f1e491c669bef133f65ff7de5819cae5c506e39d7507a6f15d5b0759d42a"} Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.367975 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.368142 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.368340 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.368556 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.368759 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.371210 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.905843 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.906848 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.907015 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.907160 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.907378 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.913605 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.942869 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38a72bf5-a9ca-409a-b52d-b10ebefc316a-kube-api-access\") pod \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\" (UID: \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\") " Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.943402 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/38a72bf5-a9ca-409a-b52d-b10ebefc316a-var-lock\") pod \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\" (UID: \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\") " Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.943451 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38a72bf5-a9ca-409a-b52d-b10ebefc316a-kubelet-dir\") pod \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\" (UID: \"38a72bf5-a9ca-409a-b52d-b10ebefc316a\") " Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.943562 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38a72bf5-a9ca-409a-b52d-b10ebefc316a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "38a72bf5-a9ca-409a-b52d-b10ebefc316a" (UID: "38a72bf5-a9ca-409a-b52d-b10ebefc316a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.943534 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38a72bf5-a9ca-409a-b52d-b10ebefc316a-var-lock" (OuterVolumeSpecName: "var-lock") pod "38a72bf5-a9ca-409a-b52d-b10ebefc316a" (UID: "38a72bf5-a9ca-409a-b52d-b10ebefc316a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.943770 4806 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/38a72bf5-a9ca-409a-b52d-b10ebefc316a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.943789 4806 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38a72bf5-a9ca-409a-b52d-b10ebefc316a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:44 crc kubenswrapper[4806]: I1127 10:25:44.969967 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38a72bf5-a9ca-409a-b52d-b10ebefc316a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "38a72bf5-a9ca-409a-b52d-b10ebefc316a" (UID: "38a72bf5-a9ca-409a-b52d-b10ebefc316a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.045445 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38a72bf5-a9ca-409a-b52d-b10ebefc316a-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.202428 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.203973 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.204836 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.205429 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.206030 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.206397 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.206714 4806 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.208453 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.247524 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.247606 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.247635 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.247650 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.247698 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.247761 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.248019 4806 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.248034 4806 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.248043 4806 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.381483 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69nx8" event={"ID":"eff0fde1-b975-41bc-95a4-080177acf5ef","Type":"ContainerStarted","Data":"66e7e4e3537f600173d0ec7bd22030ff78ac6769c2ed5f487084ee6498583eb4"} Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.383480 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh6qx" event={"ID":"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d","Type":"ContainerStarted","Data":"155ad50cfc793da40086147207ae8c331a2365531f78f50833a16d8ac3e7800e"} Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.385287 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.385975 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.386561 4806 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.387082 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.387543 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.387817 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.387892 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.388514 4806 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.388695 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61" exitCode=0 Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.388800 4806 scope.go:117] "RemoveContainer" containerID="a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.388817 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.388901 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.389167 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.389471 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.389755 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.389993 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.396003 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"38a72bf5-a9ca-409a-b52d-b10ebefc316a","Type":"ContainerDied","Data":"44121b905ac36f2b6085d34a099ee85ccac28273e2187a74ce2a54dc8a689e25"} Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.396032 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44121b905ac36f2b6085d34a099ee85ccac28273e2187a74ce2a54dc8a689e25" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.396094 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.416939 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.417127 4806 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.418423 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.418663 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.418916 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.419137 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.426501 4806 scope.go:117] "RemoveContainer" containerID="675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.431338 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.432155 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.432596 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.432827 4806 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.433759 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.433963 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.443418 4806 scope.go:117] "RemoveContainer" containerID="8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.463792 4806 scope.go:117] "RemoveContainer" containerID="d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.480031 4806 scope.go:117] "RemoveContainer" containerID="e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.498892 4806 scope.go:117] "RemoveContainer" containerID="5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.523297 4806 scope.go:117] "RemoveContainer" containerID="a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411" Nov 27 10:25:45 crc kubenswrapper[4806]: E1127 10:25:45.523739 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\": container with ID starting with a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411 not found: ID does not exist" containerID="a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.523787 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411"} err="failed to get container status \"a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\": rpc error: code = NotFound desc = could not find container \"a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411\": container with ID starting with a705170f71cd5cd70ddfb1d043e51222ad4a827335dd39b14965d0823e1a9411 not found: ID does not exist" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.523820 4806 scope.go:117] "RemoveContainer" containerID="675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280" Nov 27 10:25:45 crc kubenswrapper[4806]: E1127 10:25:45.524274 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\": container with ID starting with 675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280 not found: ID does not exist" containerID="675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.524347 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280"} err="failed to get container status \"675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\": rpc error: code = NotFound desc = could not find container \"675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280\": container with ID starting with 675d37f8e422d0e35575f11712c6855b5ed140d289a52d019cb58f73fc089280 not found: ID does not exist" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.524383 4806 scope.go:117] "RemoveContainer" containerID="8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310" Nov 27 10:25:45 crc kubenswrapper[4806]: E1127 10:25:45.524800 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\": container with ID starting with 8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310 not found: ID does not exist" containerID="8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.524837 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310"} err="failed to get container status \"8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\": rpc error: code = NotFound desc = could not find container \"8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310\": container with ID starting with 8d5fe672bdf6212cf13147250e5d8183c85cef7ed79db4db13470c1f870d7310 not found: ID does not exist" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.524882 4806 scope.go:117] "RemoveContainer" containerID="d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12" Nov 27 10:25:45 crc kubenswrapper[4806]: E1127 10:25:45.525352 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\": container with ID starting with d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12 not found: ID does not exist" containerID="d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.525379 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12"} err="failed to get container status \"d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\": rpc error: code = NotFound desc = could not find container \"d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12\": container with ID starting with d16f32b71e9485857863f450955556a72c8953161c3ba03d7b3afbfdd7237f12 not found: ID does not exist" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.525395 4806 scope.go:117] "RemoveContainer" containerID="e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61" Nov 27 10:25:45 crc kubenswrapper[4806]: E1127 10:25:45.525682 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\": container with ID starting with e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61 not found: ID does not exist" containerID="e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.525715 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61"} err="failed to get container status \"e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\": rpc error: code = NotFound desc = could not find container \"e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61\": container with ID starting with e8a6b6097414c58ba145c7ac72b7837d2efe022645bceca258cc838e1f2d6f61 not found: ID does not exist" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.525733 4806 scope.go:117] "RemoveContainer" containerID="5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28" Nov 27 10:25:45 crc kubenswrapper[4806]: E1127 10:25:45.525997 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\": container with ID starting with 5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28 not found: ID does not exist" containerID="5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28" Nov 27 10:25:45 crc kubenswrapper[4806]: I1127 10:25:45.526022 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28"} err="failed to get container status \"5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\": rpc error: code = NotFound desc = could not find container \"5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28\": container with ID starting with 5fa72c01962601e1a519777978cb9ac9d348d73a4eb950c877a1a23403462a28 not found: ID does not exist" Nov 27 10:25:46 crc kubenswrapper[4806]: I1127 10:25:46.118839 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:46 crc kubenswrapper[4806]: I1127 10:25:46.119080 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:46 crc kubenswrapper[4806]: I1127 10:25:46.119365 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:46 crc kubenswrapper[4806]: I1127 10:25:46.119653 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:46 crc kubenswrapper[4806]: I1127 10:25:46.119807 4806 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:46 crc kubenswrapper[4806]: I1127 10:25:46.119986 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:46 crc kubenswrapper[4806]: I1127 10:25:46.124763 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 27 10:25:47 crc kubenswrapper[4806]: E1127 10:25:47.702010 4806 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.246:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:47 crc kubenswrapper[4806]: I1127 10:25:47.703092 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:47 crc kubenswrapper[4806]: E1127 10:25:47.725700 4806 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:47 crc kubenswrapper[4806]: E1127 10:25:47.726152 4806 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:47 crc kubenswrapper[4806]: E1127 10:25:47.726658 4806 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:47 crc kubenswrapper[4806]: E1127 10:25:47.727129 4806 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:47 crc kubenswrapper[4806]: E1127 10:25:47.727623 4806 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:47 crc kubenswrapper[4806]: I1127 10:25:47.727654 4806 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 27 10:25:47 crc kubenswrapper[4806]: E1127 10:25:47.727841 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="200ms" Nov 27 10:25:47 crc kubenswrapper[4806]: W1127 10:25:47.735824 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-b3aa15e4acbb19c1f1ff4cefa7ef1c2b80587fe9e5c8b857c0f213fc968adc42 WatchSource:0}: Error finding container b3aa15e4acbb19c1f1ff4cefa7ef1c2b80587fe9e5c8b857c0f213fc968adc42: Status 404 returned error can't find the container with id b3aa15e4acbb19c1f1ff4cefa7ef1c2b80587fe9e5c8b857c0f213fc968adc42 Nov 27 10:25:47 crc kubenswrapper[4806]: E1127 10:25:47.929325 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="400ms" Nov 27 10:25:48 crc kubenswrapper[4806]: E1127 10:25:48.331565 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="800ms" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.417321 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"767357d89459e0619633e4816bb6394e507b3f72cbed4bb3b279ce826c1cccca"} Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.417373 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b3aa15e4acbb19c1f1ff4cefa7ef1c2b80587fe9e5c8b857c0f213fc968adc42"} Nov 27 10:25:48 crc kubenswrapper[4806]: E1127 10:25:48.418044 4806 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.246:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.418213 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.418435 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.418602 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.418996 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.419262 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.755609 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.757738 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.801571 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.801899 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.802074 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.802257 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.805065 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.805321 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.931452 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.931503 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.978673 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.979431 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.979885 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.980570 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.981146 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:48 crc kubenswrapper[4806]: I1127 10:25:48.981452 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:49 crc kubenswrapper[4806]: E1127 10:25:49.133072 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="1.6s" Nov 27 10:25:49 crc kubenswrapper[4806]: I1127 10:25:49.469322 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j54w9" Nov 27 10:25:49 crc kubenswrapper[4806]: I1127 10:25:49.470450 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:49 crc kubenswrapper[4806]: I1127 10:25:49.471013 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:49 crc kubenswrapper[4806]: I1127 10:25:49.471536 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:25:49 crc kubenswrapper[4806]: I1127 10:25:49.471662 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:49 crc kubenswrapper[4806]: I1127 10:25:49.472029 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:49 crc kubenswrapper[4806]: I1127 10:25:49.472403 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:49 crc kubenswrapper[4806]: I1127 10:25:49.472900 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:49 crc kubenswrapper[4806]: I1127 10:25:49.473269 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:49 crc kubenswrapper[4806]: I1127 10:25:49.473611 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:49 crc kubenswrapper[4806]: I1127 10:25:49.474038 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:49 crc kubenswrapper[4806]: I1127 10:25:49.474408 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:50 crc kubenswrapper[4806]: E1127 10:25:50.013784 4806 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.246:6443: connect: connection refused" event="&Event{ObjectMeta:{certified-operators-wfxvq.187bd62737861c08 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:certified-operators-wfxvq,UID:dcdfc5eb-7543-43b1-bbb5-90734e6f85a3,APIVersion:v1,ResourceVersion:29696,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 664ms (664ms including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-27 10:25:42.96928564 +0000 UTC m=+247.555876404,LastTimestamp:2025-11-27 10:25:42.96928564 +0000 UTC m=+247.555876404,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 27 10:25:50 crc kubenswrapper[4806]: E1127 10:25:50.210776 4806 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.246:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" volumeName="registry-storage" Nov 27 10:25:50 crc kubenswrapper[4806]: E1127 10:25:50.733787 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="3.2s" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.120647 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.121431 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.169641 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.170424 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.170818 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.171416 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.171665 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.171870 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.356432 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.356751 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.406567 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.407087 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.407306 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.407639 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.408251 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.408455 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.478426 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.479166 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.479580 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.480029 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.480295 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.480515 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.490054 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-69nx8" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.490755 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.491000 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.491358 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.491840 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:51 crc kubenswrapper[4806]: I1127 10:25:51.492092 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:53 crc kubenswrapper[4806]: E1127 10:25:53.934912 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.246:6443: connect: connection refused" interval="6.4s" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.116647 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.121680 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.122053 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.122543 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.123066 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.123396 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.124675 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.125195 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.125835 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.126201 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.126599 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.139909 4806 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.139954 4806 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce" Nov 27 10:25:56 crc kubenswrapper[4806]: E1127 10:25:56.140522 4806 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.141289 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:56 crc kubenswrapper[4806]: I1127 10:25:56.461764 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2bcc19ba575077197eeb2d3221d625c932b51f9924d10eed78e735f04e137879"} Nov 27 10:25:57 crc kubenswrapper[4806]: I1127 10:25:57.468513 4806 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="7575c69abb1944dfc7be7a180fed6f26814eb28066dd4e8dbed1453f3393cbf5" exitCode=0 Nov 27 10:25:57 crc kubenswrapper[4806]: I1127 10:25:57.468574 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"7575c69abb1944dfc7be7a180fed6f26814eb28066dd4e8dbed1453f3393cbf5"} Nov 27 10:25:57 crc kubenswrapper[4806]: I1127 10:25:57.468877 4806 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce" Nov 27 10:25:57 crc kubenswrapper[4806]: I1127 10:25:57.468905 4806 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce" Nov 27 10:25:57 crc kubenswrapper[4806]: E1127 10:25:57.469326 4806 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:57 crc kubenswrapper[4806]: I1127 10:25:57.469333 4806 status_manager.go:851] "Failed to get status for pod" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" pod="openshift-marketplace/community-operators-jh6qx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jh6qx\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:57 crc kubenswrapper[4806]: I1127 10:25:57.469725 4806 status_manager.go:851] "Failed to get status for pod" podUID="5e9f754d-2097-4526-9123-757c1af2a591" pod="openshift-marketplace/redhat-marketplace-j54w9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-j54w9\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:57 crc kubenswrapper[4806]: I1127 10:25:57.470308 4806 status_manager.go:851] "Failed to get status for pod" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:57 crc kubenswrapper[4806]: I1127 10:25:57.470621 4806 status_manager.go:851] "Failed to get status for pod" podUID="eff0fde1-b975-41bc-95a4-080177acf5ef" pod="openshift-marketplace/redhat-operators-69nx8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-69nx8\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:57 crc kubenswrapper[4806]: I1127 10:25:57.470919 4806 status_manager.go:851] "Failed to get status for pod" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" pod="openshift-marketplace/certified-operators-wfxvq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-wfxvq\": dial tcp 38.102.83.246:6443: connect: connection refused" Nov 27 10:25:58 crc kubenswrapper[4806]: I1127 10:25:58.488149 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 27 10:25:58 crc kubenswrapper[4806]: I1127 10:25:58.488640 4806 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8" exitCode=1 Nov 27 10:25:58 crc kubenswrapper[4806]: I1127 10:25:58.489279 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8"} Nov 27 10:25:58 crc kubenswrapper[4806]: I1127 10:25:58.494158 4806 scope.go:117] "RemoveContainer" containerID="6d2197247be02f47de4331eaf1e5cf9c762fd6c5c3032b46d92868faa8d546c8" Nov 27 10:25:58 crc kubenswrapper[4806]: I1127 10:25:58.510391 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0d5324556d96688c456a6dcb3bc516b1121553175746107e8a614e4732725e40"} Nov 27 10:25:58 crc kubenswrapper[4806]: I1127 10:25:58.510446 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4985e7e11cfaecaa9cb965c9ebbd32d27151205f029e2668e364bdc886867ad2"} Nov 27 10:25:58 crc kubenswrapper[4806]: I1127 10:25:58.510460 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"30ab72d8a33b306cf6f43c3537d1b80db410e1d9366e4ff616f76d48651527b0"} Nov 27 10:25:59 crc kubenswrapper[4806]: I1127 10:25:59.541548 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 27 10:25:59 crc kubenswrapper[4806]: I1127 10:25:59.542024 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"dd1907398dbb000bc2db0b3bd96b635fbdef08a7bfbc18d311818ed8a7f48543"} Nov 27 10:25:59 crc kubenswrapper[4806]: I1127 10:25:59.554304 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"61c47dc3b0d956131d11a7f88d1874b6e214404d07bcf083bde39576b758e13f"} Nov 27 10:25:59 crc kubenswrapper[4806]: I1127 10:25:59.554351 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"28708e77b6346a2d0f0b31925a66c8b0e1e798d507aeeae8ee5920ac7795a6d9"} Nov 27 10:25:59 crc kubenswrapper[4806]: I1127 10:25:59.554817 4806 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce" Nov 27 10:25:59 crc kubenswrapper[4806]: I1127 10:25:59.554820 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:25:59 crc kubenswrapper[4806]: I1127 10:25:59.554842 4806 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce" Nov 27 10:26:01 crc kubenswrapper[4806]: I1127 10:26:01.141471 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:26:01 crc kubenswrapper[4806]: I1127 10:26:01.142132 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:26:01 crc kubenswrapper[4806]: I1127 10:26:01.149640 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:26:02 crc kubenswrapper[4806]: I1127 10:26:02.074521 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:26:02 crc kubenswrapper[4806]: I1127 10:26:02.509999 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:26:02 crc kubenswrapper[4806]: I1127 10:26:02.514299 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:26:04 crc kubenswrapper[4806]: I1127 10:26:04.744476 4806 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:26:05 crc kubenswrapper[4806]: I1127 10:26:05.588497 4806 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce" Nov 27 10:26:05 crc kubenswrapper[4806]: I1127 10:26:05.588539 4806 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce" Nov 27 10:26:05 crc kubenswrapper[4806]: I1127 10:26:05.593447 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:26:06 crc kubenswrapper[4806]: I1127 10:26:06.151006 4806 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="c111ea9e-d9d5-4130-addf-179cbbda0b7d" Nov 27 10:26:06 crc kubenswrapper[4806]: I1127 10:26:06.595076 4806 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce" Nov 27 10:26:06 crc kubenswrapper[4806]: I1127 10:26:06.595124 4806 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="098c4ff4-2bcc-413f-afe7-5dc2f41cd4ce" Nov 27 10:26:06 crc kubenswrapper[4806]: I1127 10:26:06.598849 4806 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="c111ea9e-d9d5-4130-addf-179cbbda0b7d" Nov 27 10:26:12 crc kubenswrapper[4806]: I1127 10:26:12.078408 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 27 10:26:14 crc kubenswrapper[4806]: I1127 10:26:14.694426 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 27 10:26:15 crc kubenswrapper[4806]: I1127 10:26:15.036199 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 27 10:26:15 crc kubenswrapper[4806]: I1127 10:26:15.107163 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 27 10:26:15 crc kubenswrapper[4806]: I1127 10:26:15.841601 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 27 10:26:16 crc kubenswrapper[4806]: I1127 10:26:16.001783 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 27 10:26:16 crc kubenswrapper[4806]: I1127 10:26:16.052755 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 27 10:26:16 crc kubenswrapper[4806]: I1127 10:26:16.356135 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 27 10:26:16 crc kubenswrapper[4806]: I1127 10:26:16.614115 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 27 10:26:16 crc kubenswrapper[4806]: I1127 10:26:16.694121 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 27 10:26:16 crc kubenswrapper[4806]: I1127 10:26:16.831804 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 27 10:26:16 crc kubenswrapper[4806]: I1127 10:26:16.914446 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 27 10:26:17 crc kubenswrapper[4806]: I1127 10:26:17.087394 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 27 10:26:17 crc kubenswrapper[4806]: I1127 10:26:17.095592 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 27 10:26:17 crc kubenswrapper[4806]: I1127 10:26:17.136222 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 27 10:26:17 crc kubenswrapper[4806]: I1127 10:26:17.213665 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 27 10:26:17 crc kubenswrapper[4806]: I1127 10:26:17.337810 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 27 10:26:17 crc kubenswrapper[4806]: I1127 10:26:17.486285 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 27 10:26:17 crc kubenswrapper[4806]: I1127 10:26:17.775963 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 27 10:26:17 crc kubenswrapper[4806]: I1127 10:26:17.821700 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 27 10:26:17 crc kubenswrapper[4806]: I1127 10:26:17.905269 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.167822 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.233382 4806 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.263767 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.278769 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.301207 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.416351 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.517873 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.521846 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.540353 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.553375 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.570355 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.717878 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.821815 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.889957 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.890637 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.909869 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.929814 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.930668 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.959108 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.982632 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 27 10:26:18 crc kubenswrapper[4806]: I1127 10:26:18.993741 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.015562 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.101033 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.197771 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.263131 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.341190 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.421325 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.448097 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.479008 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.587783 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.598646 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.648279 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.649955 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.884583 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.892693 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.908677 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 27 10:26:19 crc kubenswrapper[4806]: I1127 10:26:19.931891 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.054732 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.058935 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.143047 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.221580 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.231281 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.263751 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.323939 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.325439 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.337137 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.372131 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.471575 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.513618 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.525889 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.555800 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.622304 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.655976 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.703982 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.748466 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.836101 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 27 10:26:20 crc kubenswrapper[4806]: I1127 10:26:20.992731 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.005728 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.030245 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.055255 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.081895 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.097179 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.120651 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.133344 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.134785 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.144760 4806 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.161343 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.192418 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.217933 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.251630 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.322487 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.323763 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.326103 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.488968 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.510101 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.542501 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.555784 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.584041 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.721872 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.835811 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.864547 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.915107 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.940106 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.961923 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 27 10:26:21 crc kubenswrapper[4806]: I1127 10:26:21.968092 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 27 10:26:22 crc kubenswrapper[4806]: I1127 10:26:22.072461 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 27 10:26:22 crc kubenswrapper[4806]: I1127 10:26:22.083377 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 27 10:26:22 crc kubenswrapper[4806]: I1127 10:26:22.180139 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 27 10:26:22 crc kubenswrapper[4806]: I1127 10:26:22.285786 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 27 10:26:22 crc kubenswrapper[4806]: I1127 10:26:22.372726 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 27 10:26:22 crc kubenswrapper[4806]: I1127 10:26:22.375714 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 27 10:26:22 crc kubenswrapper[4806]: I1127 10:26:22.437211 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 27 10:26:22 crc kubenswrapper[4806]: I1127 10:26:22.512889 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 27 10:26:22 crc kubenswrapper[4806]: I1127 10:26:22.639139 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 27 10:26:22 crc kubenswrapper[4806]: I1127 10:26:22.647979 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 27 10:26:22 crc kubenswrapper[4806]: I1127 10:26:22.655988 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 27 10:26:22 crc kubenswrapper[4806]: I1127 10:26:22.665932 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 27 10:26:23 crc kubenswrapper[4806]: I1127 10:26:23.037213 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 27 10:26:23 crc kubenswrapper[4806]: I1127 10:26:23.265032 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 27 10:26:23 crc kubenswrapper[4806]: I1127 10:26:23.265630 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 27 10:26:23 crc kubenswrapper[4806]: I1127 10:26:23.283347 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 27 10:26:23 crc kubenswrapper[4806]: I1127 10:26:23.284904 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 27 10:26:23 crc kubenswrapper[4806]: I1127 10:26:23.328461 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 27 10:26:23 crc kubenswrapper[4806]: I1127 10:26:23.432534 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 27 10:26:23 crc kubenswrapper[4806]: I1127 10:26:23.511126 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 27 10:26:23 crc kubenswrapper[4806]: I1127 10:26:23.585305 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 27 10:26:23 crc kubenswrapper[4806]: I1127 10:26:23.596808 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 27 10:26:23 crc kubenswrapper[4806]: I1127 10:26:23.637801 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 27 10:26:23 crc kubenswrapper[4806]: I1127 10:26:23.644273 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 27 10:26:23 crc kubenswrapper[4806]: I1127 10:26:23.870343 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.021408 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.138673 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.152156 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.176673 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.183951 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.222223 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.268087 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.307685 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.336279 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.392502 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.431511 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.482399 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.608585 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.637986 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.697598 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.858141 4806 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.888284 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 27 10:26:24 crc kubenswrapper[4806]: I1127 10:26:24.890868 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.054786 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.184081 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.197927 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.222205 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.226597 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.242934 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.260349 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.265062 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.292073 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.338578 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.386017 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.432861 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.444682 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.483873 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.530173 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.553241 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.600657 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.617880 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.694866 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.711140 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.836001 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.891883 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 27 10:26:25 crc kubenswrapper[4806]: I1127 10:26:25.892782 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.030529 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.094391 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.107665 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.169731 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.181740 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.264392 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.307489 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.349134 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.497970 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.565706 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.575790 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.645364 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.654252 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.666061 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.686276 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.696918 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.750551 4806 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.751206 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wfxvq" podStartSLOduration=46.014186364 podStartE2EDuration="48.751192289s" podCreationTimestamp="2025-11-27 10:25:38 +0000 UTC" firstStartedPulling="2025-11-27 10:25:40.232265445 +0000 UTC m=+244.818856209" lastFinishedPulling="2025-11-27 10:25:42.96927137 +0000 UTC m=+247.555862134" observedRunningTime="2025-11-27 10:26:04.837948253 +0000 UTC m=+269.424539007" watchObservedRunningTime="2025-11-27 10:26:26.751192289 +0000 UTC m=+291.337783053" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.751653 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j54w9" podStartSLOduration=45.978331988 podStartE2EDuration="48.751649622s" podCreationTimestamp="2025-11-27 10:25:38 +0000 UTC" firstStartedPulling="2025-11-27 10:25:40.227022895 +0000 UTC m=+244.813613679" lastFinishedPulling="2025-11-27 10:25:43.000340539 +0000 UTC m=+247.586931313" observedRunningTime="2025-11-27 10:26:04.896557745 +0000 UTC m=+269.483148509" watchObservedRunningTime="2025-11-27 10:26:26.751649622 +0000 UTC m=+291.338240386" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.752843 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-69nx8" podStartSLOduration=44.063689084 podStartE2EDuration="46.752839697s" podCreationTimestamp="2025-11-27 10:25:40 +0000 UTC" firstStartedPulling="2025-11-27 10:25:42.311849387 +0000 UTC m=+246.898440151" lastFinishedPulling="2025-11-27 10:25:45.00099999 +0000 UTC m=+249.587590764" observedRunningTime="2025-11-27 10:26:04.802264673 +0000 UTC m=+269.388855447" watchObservedRunningTime="2025-11-27 10:26:26.752839697 +0000 UTC m=+291.339430461" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.753675 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jh6qx" podStartSLOduration=44.202922911 podStartE2EDuration="46.75367153s" podCreationTimestamp="2025-11-27 10:25:40 +0000 UTC" firstStartedPulling="2025-11-27 10:25:42.320046701 +0000 UTC m=+246.906637475" lastFinishedPulling="2025-11-27 10:25:44.87079534 +0000 UTC m=+249.457386094" observedRunningTime="2025-11-27 10:26:04.871166752 +0000 UTC m=+269.457757516" watchObservedRunningTime="2025-11-27 10:26:26.75367153 +0000 UTC m=+291.340262294" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.754475 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.754513 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.759258 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.777847 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=22.777821008 podStartE2EDuration="22.777821008s" podCreationTimestamp="2025-11-27 10:26:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:26:26.772049561 +0000 UTC m=+291.358640345" watchObservedRunningTime="2025-11-27 10:26:26.777821008 +0000 UTC m=+291.364411772" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.789523 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.868192 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.869271 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.925858 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 27 10:26:26 crc kubenswrapper[4806]: I1127 10:26:26.996111 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 27 10:26:27 crc kubenswrapper[4806]: I1127 10:26:27.006360 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 27 10:26:27 crc kubenswrapper[4806]: I1127 10:26:27.049675 4806 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 27 10:26:27 crc kubenswrapper[4806]: I1127 10:26:27.255853 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 27 10:26:27 crc kubenswrapper[4806]: I1127 10:26:27.287972 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 27 10:26:27 crc kubenswrapper[4806]: I1127 10:26:27.486969 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 27 10:26:27 crc kubenswrapper[4806]: I1127 10:26:27.535873 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 27 10:26:27 crc kubenswrapper[4806]: I1127 10:26:27.601950 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 27 10:26:27 crc kubenswrapper[4806]: I1127 10:26:27.675736 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 27 10:26:27 crc kubenswrapper[4806]: I1127 10:26:27.781337 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 27 10:26:27 crc kubenswrapper[4806]: I1127 10:26:27.786912 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 27 10:26:27 crc kubenswrapper[4806]: I1127 10:26:27.950934 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 27 10:26:27 crc kubenswrapper[4806]: I1127 10:26:27.977036 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.122191 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.256921 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.264554 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.287095 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.297510 4806 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.374473 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.408194 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.496440 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.549704 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.666150 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.719636 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.772987 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.841118 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 27 10:26:28 crc kubenswrapper[4806]: I1127 10:26:28.892749 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.142515 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.185859 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.239215 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.275553 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.306463 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.438807 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.479125 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.508173 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.540939 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.560258 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.590924 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.652494 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.708001 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.763463 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.791210 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.852055 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.935474 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 27 10:26:29 crc kubenswrapper[4806]: I1127 10:26:29.988823 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 27 10:26:30 crc kubenswrapper[4806]: I1127 10:26:30.225313 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 27 10:26:30 crc kubenswrapper[4806]: I1127 10:26:30.237539 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 27 10:26:30 crc kubenswrapper[4806]: I1127 10:26:30.267040 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 27 10:26:30 crc kubenswrapper[4806]: I1127 10:26:30.275962 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 27 10:26:30 crc kubenswrapper[4806]: I1127 10:26:30.287078 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 27 10:26:30 crc kubenswrapper[4806]: I1127 10:26:30.294056 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 27 10:26:30 crc kubenswrapper[4806]: I1127 10:26:30.436733 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 27 10:26:30 crc kubenswrapper[4806]: I1127 10:26:30.500154 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 27 10:26:30 crc kubenswrapper[4806]: I1127 10:26:30.668707 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 27 10:26:30 crc kubenswrapper[4806]: I1127 10:26:30.679341 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 27 10:26:30 crc kubenswrapper[4806]: I1127 10:26:30.744620 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 27 10:26:30 crc kubenswrapper[4806]: I1127 10:26:30.805223 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 27 10:26:31 crc kubenswrapper[4806]: I1127 10:26:31.036058 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 27 10:26:31 crc kubenswrapper[4806]: I1127 10:26:31.920436 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 27 10:26:37 crc kubenswrapper[4806]: I1127 10:26:37.501966 4806 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 27 10:26:37 crc kubenswrapper[4806]: I1127 10:26:37.503178 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://767357d89459e0619633e4816bb6394e507b3f72cbed4bb3b279ce826c1cccca" gracePeriod=5 Nov 27 10:26:42 crc kubenswrapper[4806]: I1127 10:26:42.818261 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 27 10:26:42 crc kubenswrapper[4806]: I1127 10:26:42.818709 4806 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="767357d89459e0619633e4816bb6394e507b3f72cbed4bb3b279ce826c1cccca" exitCode=137 Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.105000 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.105126 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.211524 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.211670 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.211711 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.211769 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.211813 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.211850 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.211885 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.211892 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.211988 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.213511 4806 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.213538 4806 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.213551 4806 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.213567 4806 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.233455 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.314695 4806 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.825071 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.825457 4806 scope.go:117] "RemoveContainer" containerID="767357d89459e0619633e4816bb6394e507b3f72cbed4bb3b279ce826c1cccca" Nov 27 10:26:43 crc kubenswrapper[4806]: I1127 10:26:43.825503 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 27 10:26:44 crc kubenswrapper[4806]: I1127 10:26:44.125490 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 27 10:27:13 crc kubenswrapper[4806]: I1127 10:27:13.300321 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.226508 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-vvcjg"] Nov 27 10:27:14 crc kubenswrapper[4806]: E1127 10:27:14.227113 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" containerName="installer" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.227135 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" containerName="installer" Nov 27 10:27:14 crc kubenswrapper[4806]: E1127 10:27:14.227151 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.227159 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.227301 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="38a72bf5-a9ca-409a-b52d-b10ebefc316a" containerName="installer" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.227321 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.227844 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.275563 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-vvcjg"] Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.386623 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxgm2\" (UniqueName: \"kubernetes.io/projected/7039cbc3-5594-4e86-b4d9-28324b97a8d5-kube-api-access-jxgm2\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.386683 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7039cbc3-5594-4e86-b4d9-28324b97a8d5-registry-certificates\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.386733 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.386853 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7039cbc3-5594-4e86-b4d9-28324b97a8d5-trusted-ca\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.386926 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7039cbc3-5594-4e86-b4d9-28324b97a8d5-bound-sa-token\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.386952 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7039cbc3-5594-4e86-b4d9-28324b97a8d5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.386978 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7039cbc3-5594-4e86-b4d9-28324b97a8d5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.387041 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7039cbc3-5594-4e86-b4d9-28324b97a8d5-registry-tls\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.412624 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.451128 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.451193 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.488297 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7039cbc3-5594-4e86-b4d9-28324b97a8d5-registry-tls\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.488350 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxgm2\" (UniqueName: \"kubernetes.io/projected/7039cbc3-5594-4e86-b4d9-28324b97a8d5-kube-api-access-jxgm2\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.488390 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7039cbc3-5594-4e86-b4d9-28324b97a8d5-registry-certificates\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.488440 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7039cbc3-5594-4e86-b4d9-28324b97a8d5-trusted-ca\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.488459 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7039cbc3-5594-4e86-b4d9-28324b97a8d5-bound-sa-token\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.488477 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7039cbc3-5594-4e86-b4d9-28324b97a8d5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.488508 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7039cbc3-5594-4e86-b4d9-28324b97a8d5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.489368 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7039cbc3-5594-4e86-b4d9-28324b97a8d5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.490194 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7039cbc3-5594-4e86-b4d9-28324b97a8d5-registry-certificates\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.491804 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7039cbc3-5594-4e86-b4d9-28324b97a8d5-trusted-ca\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.493588 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7039cbc3-5594-4e86-b4d9-28324b97a8d5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.493586 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7039cbc3-5594-4e86-b4d9-28324b97a8d5-registry-tls\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.507090 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxgm2\" (UniqueName: \"kubernetes.io/projected/7039cbc3-5594-4e86-b4d9-28324b97a8d5-kube-api-access-jxgm2\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.507211 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7039cbc3-5594-4e86-b4d9-28324b97a8d5-bound-sa-token\") pod \"image-registry-66df7c8f76-vvcjg\" (UID: \"7039cbc3-5594-4e86-b4d9-28324b97a8d5\") " pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.545507 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.779052 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-vvcjg"] Nov 27 10:27:14 crc kubenswrapper[4806]: W1127 10:27:14.785680 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7039cbc3_5594_4e86_b4d9_28324b97a8d5.slice/crio-5a6ebd9946788e0680dc6c0ba65f76e69b257d1a4f704896471ae682af2a00d8 WatchSource:0}: Error finding container 5a6ebd9946788e0680dc6c0ba65f76e69b257d1a4f704896471ae682af2a00d8: Status 404 returned error can't find the container with id 5a6ebd9946788e0680dc6c0ba65f76e69b257d1a4f704896471ae682af2a00d8 Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.998904 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" event={"ID":"7039cbc3-5594-4e86-b4d9-28324b97a8d5","Type":"ContainerStarted","Data":"e23aaf2fbeb720dfff0d21f0e5719e9dd8242bbf0e7bb07f3fe5795d4ee0e003"} Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.999173 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" event={"ID":"7039cbc3-5594-4e86-b4d9-28324b97a8d5","Type":"ContainerStarted","Data":"5a6ebd9946788e0680dc6c0ba65f76e69b257d1a4f704896471ae682af2a00d8"} Nov 27 10:27:14 crc kubenswrapper[4806]: I1127 10:27:14.999194 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:15 crc kubenswrapper[4806]: I1127 10:27:15.025161 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" podStartSLOduration=1.025138124 podStartE2EDuration="1.025138124s" podCreationTimestamp="2025-11-27 10:27:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:27:15.017681587 +0000 UTC m=+339.604272351" watchObservedRunningTime="2025-11-27 10:27:15.025138124 +0000 UTC m=+339.611728898" Nov 27 10:27:34 crc kubenswrapper[4806]: I1127 10:27:34.550652 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-vvcjg" Nov 27 10:27:34 crc kubenswrapper[4806]: I1127 10:27:34.607396 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kmbdt"] Nov 27 10:27:44 crc kubenswrapper[4806]: I1127 10:27:44.451178 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:27:44 crc kubenswrapper[4806]: I1127 10:27:44.452029 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:27:59 crc kubenswrapper[4806]: I1127 10:27:59.662356 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" podUID="e2a6f027-898d-4a60-aa19-00b6f54d2aac" containerName="registry" containerID="cri-o://f830241c193cfa12475072329c53166088006fa0850175b330859d3b71bf9ead" gracePeriod=30 Nov 27 10:27:59 crc kubenswrapper[4806]: I1127 10:27:59.990224 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.088871 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2a6f027-898d-4a60-aa19-00b6f54d2aac-ca-trust-extracted\") pod \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.088978 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2a6f027-898d-4a60-aa19-00b6f54d2aac-installation-pull-secrets\") pod \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.089024 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2a6f027-898d-4a60-aa19-00b6f54d2aac-trusted-ca\") pod \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.089208 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.089312 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2a6f027-898d-4a60-aa19-00b6f54d2aac-registry-certificates\") pod \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.089367 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-bound-sa-token\") pod \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.089433 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-registry-tls\") pod \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.089473 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrlcf\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-kube-api-access-lrlcf\") pod \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\" (UID: \"e2a6f027-898d-4a60-aa19-00b6f54d2aac\") " Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.089655 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2a6f027-898d-4a60-aa19-00b6f54d2aac-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "e2a6f027-898d-4a60-aa19-00b6f54d2aac" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.089803 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2a6f027-898d-4a60-aa19-00b6f54d2aac-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "e2a6f027-898d-4a60-aa19-00b6f54d2aac" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.094502 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "e2a6f027-898d-4a60-aa19-00b6f54d2aac" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.095503 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2a6f027-898d-4a60-aa19-00b6f54d2aac-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "e2a6f027-898d-4a60-aa19-00b6f54d2aac" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.101444 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "e2a6f027-898d-4a60-aa19-00b6f54d2aac" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.103255 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-kube-api-access-lrlcf" (OuterVolumeSpecName: "kube-api-access-lrlcf") pod "e2a6f027-898d-4a60-aa19-00b6f54d2aac" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac"). InnerVolumeSpecName "kube-api-access-lrlcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.108457 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "e2a6f027-898d-4a60-aa19-00b6f54d2aac" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.116545 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2a6f027-898d-4a60-aa19-00b6f54d2aac-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "e2a6f027-898d-4a60-aa19-00b6f54d2aac" (UID: "e2a6f027-898d-4a60-aa19-00b6f54d2aac"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.190522 4806 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.190574 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrlcf\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-kube-api-access-lrlcf\") on node \"crc\" DevicePath \"\"" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.190587 4806 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2a6f027-898d-4a60-aa19-00b6f54d2aac-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.190596 4806 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2a6f027-898d-4a60-aa19-00b6f54d2aac-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.190605 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2a6f027-898d-4a60-aa19-00b6f54d2aac-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.190612 4806 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2a6f027-898d-4a60-aa19-00b6f54d2aac-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.190620 4806 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2a6f027-898d-4a60-aa19-00b6f54d2aac-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.275681 4806 generic.go:334] "Generic (PLEG): container finished" podID="e2a6f027-898d-4a60-aa19-00b6f54d2aac" containerID="f830241c193cfa12475072329c53166088006fa0850175b330859d3b71bf9ead" exitCode=0 Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.275723 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" event={"ID":"e2a6f027-898d-4a60-aa19-00b6f54d2aac","Type":"ContainerDied","Data":"f830241c193cfa12475072329c53166088006fa0850175b330859d3b71bf9ead"} Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.275746 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" event={"ID":"e2a6f027-898d-4a60-aa19-00b6f54d2aac","Type":"ContainerDied","Data":"27b5372bda575891bd1567d47d1f0de3acc3b266792acb3269aba0ab7af15df2"} Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.275762 4806 scope.go:117] "RemoveContainer" containerID="f830241c193cfa12475072329c53166088006fa0850175b330859d3b71bf9ead" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.275767 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kmbdt" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.298503 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kmbdt"] Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.300621 4806 scope.go:117] "RemoveContainer" containerID="f830241c193cfa12475072329c53166088006fa0850175b330859d3b71bf9ead" Nov 27 10:28:00 crc kubenswrapper[4806]: E1127 10:28:00.301832 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f830241c193cfa12475072329c53166088006fa0850175b330859d3b71bf9ead\": container with ID starting with f830241c193cfa12475072329c53166088006fa0850175b330859d3b71bf9ead not found: ID does not exist" containerID="f830241c193cfa12475072329c53166088006fa0850175b330859d3b71bf9ead" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.301873 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f830241c193cfa12475072329c53166088006fa0850175b330859d3b71bf9ead"} err="failed to get container status \"f830241c193cfa12475072329c53166088006fa0850175b330859d3b71bf9ead\": rpc error: code = NotFound desc = could not find container \"f830241c193cfa12475072329c53166088006fa0850175b330859d3b71bf9ead\": container with ID starting with f830241c193cfa12475072329c53166088006fa0850175b330859d3b71bf9ead not found: ID does not exist" Nov 27 10:28:00 crc kubenswrapper[4806]: I1127 10:28:00.303569 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kmbdt"] Nov 27 10:28:02 crc kubenswrapper[4806]: I1127 10:28:02.127060 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2a6f027-898d-4a60-aa19-00b6f54d2aac" path="/var/lib/kubelet/pods/e2a6f027-898d-4a60-aa19-00b6f54d2aac/volumes" Nov 27 10:28:14 crc kubenswrapper[4806]: I1127 10:28:14.450698 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:28:14 crc kubenswrapper[4806]: I1127 10:28:14.451191 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:28:14 crc kubenswrapper[4806]: I1127 10:28:14.451259 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:28:14 crc kubenswrapper[4806]: I1127 10:28:14.451823 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0f2788d6ded82e196b1f288174c14861105e5b2e48bf181245fe8449fb3b1acb"} pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 10:28:14 crc kubenswrapper[4806]: I1127 10:28:14.451876 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" containerID="cri-o://0f2788d6ded82e196b1f288174c14861105e5b2e48bf181245fe8449fb3b1acb" gracePeriod=600 Nov 27 10:28:15 crc kubenswrapper[4806]: I1127 10:28:15.362420 4806 generic.go:334] "Generic (PLEG): container finished" podID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerID="0f2788d6ded82e196b1f288174c14861105e5b2e48bf181245fe8449fb3b1acb" exitCode=0 Nov 27 10:28:15 crc kubenswrapper[4806]: I1127 10:28:15.362957 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerDied","Data":"0f2788d6ded82e196b1f288174c14861105e5b2e48bf181245fe8449fb3b1acb"} Nov 27 10:28:15 crc kubenswrapper[4806]: I1127 10:28:15.362990 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"27aa6f365485f368ff18494a5978d93bb6e961422bea73da5d7f2d1830e740c0"} Nov 27 10:28:15 crc kubenswrapper[4806]: I1127 10:28:15.363018 4806 scope.go:117] "RemoveContainer" containerID="7bfe899213eeaee6911474e3cd8e6b122d0642b6acf9bbf4abe48346f84d4218" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.191276 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc"] Nov 27 10:30:00 crc kubenswrapper[4806]: E1127 10:30:00.192048 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2a6f027-898d-4a60-aa19-00b6f54d2aac" containerName="registry" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.192066 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2a6f027-898d-4a60-aa19-00b6f54d2aac" containerName="registry" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.192201 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2a6f027-898d-4a60-aa19-00b6f54d2aac" containerName="registry" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.192744 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.194898 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.195426 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.206765 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc"] Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.297493 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/34f02fc8-06ef-418e-9830-f6f05cd6068a-config-volume\") pod \"collect-profiles-29403990-gxqkc\" (UID: \"34f02fc8-06ef-418e-9830-f6f05cd6068a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.297549 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw9kx\" (UniqueName: \"kubernetes.io/projected/34f02fc8-06ef-418e-9830-f6f05cd6068a-kube-api-access-qw9kx\") pod \"collect-profiles-29403990-gxqkc\" (UID: \"34f02fc8-06ef-418e-9830-f6f05cd6068a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.297682 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/34f02fc8-06ef-418e-9830-f6f05cd6068a-secret-volume\") pod \"collect-profiles-29403990-gxqkc\" (UID: \"34f02fc8-06ef-418e-9830-f6f05cd6068a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.398917 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/34f02fc8-06ef-418e-9830-f6f05cd6068a-secret-volume\") pod \"collect-profiles-29403990-gxqkc\" (UID: \"34f02fc8-06ef-418e-9830-f6f05cd6068a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.398971 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/34f02fc8-06ef-418e-9830-f6f05cd6068a-config-volume\") pod \"collect-profiles-29403990-gxqkc\" (UID: \"34f02fc8-06ef-418e-9830-f6f05cd6068a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.398987 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw9kx\" (UniqueName: \"kubernetes.io/projected/34f02fc8-06ef-418e-9830-f6f05cd6068a-kube-api-access-qw9kx\") pod \"collect-profiles-29403990-gxqkc\" (UID: \"34f02fc8-06ef-418e-9830-f6f05cd6068a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.400817 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/34f02fc8-06ef-418e-9830-f6f05cd6068a-config-volume\") pod \"collect-profiles-29403990-gxqkc\" (UID: \"34f02fc8-06ef-418e-9830-f6f05cd6068a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.409645 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/34f02fc8-06ef-418e-9830-f6f05cd6068a-secret-volume\") pod \"collect-profiles-29403990-gxqkc\" (UID: \"34f02fc8-06ef-418e-9830-f6f05cd6068a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.415685 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw9kx\" (UniqueName: \"kubernetes.io/projected/34f02fc8-06ef-418e-9830-f6f05cd6068a-kube-api-access-qw9kx\") pod \"collect-profiles-29403990-gxqkc\" (UID: \"34f02fc8-06ef-418e-9830-f6f05cd6068a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.515646 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" Nov 27 10:30:00 crc kubenswrapper[4806]: I1127 10:30:00.729794 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc"] Nov 27 10:30:01 crc kubenswrapper[4806]: I1127 10:30:01.203855 4806 generic.go:334] "Generic (PLEG): container finished" podID="34f02fc8-06ef-418e-9830-f6f05cd6068a" containerID="a0935a79653dd0e3955523f1f47aac259f1db3ce48451c7ef29dbfb11c19f52c" exitCode=0 Nov 27 10:30:01 crc kubenswrapper[4806]: I1127 10:30:01.203985 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" event={"ID":"34f02fc8-06ef-418e-9830-f6f05cd6068a","Type":"ContainerDied","Data":"a0935a79653dd0e3955523f1f47aac259f1db3ce48451c7ef29dbfb11c19f52c"} Nov 27 10:30:01 crc kubenswrapper[4806]: I1127 10:30:01.204272 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" event={"ID":"34f02fc8-06ef-418e-9830-f6f05cd6068a","Type":"ContainerStarted","Data":"1b854bf3e5b84025616e84c5db84f9700fbbdaac0e42470b4549e16094705eea"} Nov 27 10:30:02 crc kubenswrapper[4806]: I1127 10:30:02.432281 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" Nov 27 10:30:02 crc kubenswrapper[4806]: I1127 10:30:02.626848 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/34f02fc8-06ef-418e-9830-f6f05cd6068a-config-volume\") pod \"34f02fc8-06ef-418e-9830-f6f05cd6068a\" (UID: \"34f02fc8-06ef-418e-9830-f6f05cd6068a\") " Nov 27 10:30:02 crc kubenswrapper[4806]: I1127 10:30:02.626966 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/34f02fc8-06ef-418e-9830-f6f05cd6068a-secret-volume\") pod \"34f02fc8-06ef-418e-9830-f6f05cd6068a\" (UID: \"34f02fc8-06ef-418e-9830-f6f05cd6068a\") " Nov 27 10:30:02 crc kubenswrapper[4806]: I1127 10:30:02.626987 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw9kx\" (UniqueName: \"kubernetes.io/projected/34f02fc8-06ef-418e-9830-f6f05cd6068a-kube-api-access-qw9kx\") pod \"34f02fc8-06ef-418e-9830-f6f05cd6068a\" (UID: \"34f02fc8-06ef-418e-9830-f6f05cd6068a\") " Nov 27 10:30:02 crc kubenswrapper[4806]: I1127 10:30:02.627912 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34f02fc8-06ef-418e-9830-f6f05cd6068a-config-volume" (OuterVolumeSpecName: "config-volume") pod "34f02fc8-06ef-418e-9830-f6f05cd6068a" (UID: "34f02fc8-06ef-418e-9830-f6f05cd6068a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:30:02 crc kubenswrapper[4806]: I1127 10:30:02.634369 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34f02fc8-06ef-418e-9830-f6f05cd6068a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "34f02fc8-06ef-418e-9830-f6f05cd6068a" (UID: "34f02fc8-06ef-418e-9830-f6f05cd6068a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:30:02 crc kubenswrapper[4806]: I1127 10:30:02.634482 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34f02fc8-06ef-418e-9830-f6f05cd6068a-kube-api-access-qw9kx" (OuterVolumeSpecName: "kube-api-access-qw9kx") pod "34f02fc8-06ef-418e-9830-f6f05cd6068a" (UID: "34f02fc8-06ef-418e-9830-f6f05cd6068a"). InnerVolumeSpecName "kube-api-access-qw9kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:30:02 crc kubenswrapper[4806]: I1127 10:30:02.728379 4806 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/34f02fc8-06ef-418e-9830-f6f05cd6068a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:02 crc kubenswrapper[4806]: I1127 10:30:02.728439 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw9kx\" (UniqueName: \"kubernetes.io/projected/34f02fc8-06ef-418e-9830-f6f05cd6068a-kube-api-access-qw9kx\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:02 crc kubenswrapper[4806]: I1127 10:30:02.728460 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/34f02fc8-06ef-418e-9830-f6f05cd6068a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:03 crc kubenswrapper[4806]: I1127 10:30:03.213690 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" event={"ID":"34f02fc8-06ef-418e-9830-f6f05cd6068a","Type":"ContainerDied","Data":"1b854bf3e5b84025616e84c5db84f9700fbbdaac0e42470b4549e16094705eea"} Nov 27 10:30:03 crc kubenswrapper[4806]: I1127 10:30:03.213733 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b854bf3e5b84025616e84c5db84f9700fbbdaac0e42470b4549e16094705eea" Nov 27 10:30:03 crc kubenswrapper[4806]: I1127 10:30:03.213789 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29403990-gxqkc" Nov 27 10:30:14 crc kubenswrapper[4806]: I1127 10:30:14.450720 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:30:14 crc kubenswrapper[4806]: I1127 10:30:14.451166 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:30:44 crc kubenswrapper[4806]: I1127 10:30:44.450317 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:30:44 crc kubenswrapper[4806]: I1127 10:30:44.451186 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.095493 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-6dclx"] Nov 27 10:30:45 crc kubenswrapper[4806]: E1127 10:30:45.095846 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f02fc8-06ef-418e-9830-f6f05cd6068a" containerName="collect-profiles" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.095862 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f02fc8-06ef-418e-9830-f6f05cd6068a" containerName="collect-profiles" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.096001 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="34f02fc8-06ef-418e-9830-f6f05cd6068a" containerName="collect-profiles" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.096596 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-6dclx" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.099291 4806 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-7z7n8" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.101331 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.102186 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.126140 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-6dclx"] Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.137549 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-ptpz9"] Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.138425 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-ptpz9" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.140538 4806 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-qs42q" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.141644 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-pvxdj"] Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.142367 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-pvxdj" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.145890 4806 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-qfgsh" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.160510 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcztx\" (UniqueName: \"kubernetes.io/projected/42836747-f4cd-43a2-a690-b17989a8d2c1-kube-api-access-zcztx\") pod \"cert-manager-cainjector-7f985d654d-6dclx\" (UID: \"42836747-f4cd-43a2-a690-b17989a8d2c1\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-6dclx" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.163947 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-ptpz9"] Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.206470 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-pvxdj"] Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.261617 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcztx\" (UniqueName: \"kubernetes.io/projected/42836747-f4cd-43a2-a690-b17989a8d2c1-kube-api-access-zcztx\") pod \"cert-manager-cainjector-7f985d654d-6dclx\" (UID: \"42836747-f4cd-43a2-a690-b17989a8d2c1\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-6dclx" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.261920 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ncn2\" (UniqueName: \"kubernetes.io/projected/2e8a774f-7876-4f56-94e6-c559397129c1-kube-api-access-7ncn2\") pod \"cert-manager-5b446d88c5-pvxdj\" (UID: \"2e8a774f-7876-4f56-94e6-c559397129c1\") " pod="cert-manager/cert-manager-5b446d88c5-pvxdj" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.262036 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nss7r\" (UniqueName: \"kubernetes.io/projected/9f1cba27-3a94-461f-bee4-1dde7d31f9a6-kube-api-access-nss7r\") pod \"cert-manager-webhook-5655c58dd6-ptpz9\" (UID: \"9f1cba27-3a94-461f-bee4-1dde7d31f9a6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-ptpz9" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.284684 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcztx\" (UniqueName: \"kubernetes.io/projected/42836747-f4cd-43a2-a690-b17989a8d2c1-kube-api-access-zcztx\") pod \"cert-manager-cainjector-7f985d654d-6dclx\" (UID: \"42836747-f4cd-43a2-a690-b17989a8d2c1\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-6dclx" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.363803 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ncn2\" (UniqueName: \"kubernetes.io/projected/2e8a774f-7876-4f56-94e6-c559397129c1-kube-api-access-7ncn2\") pod \"cert-manager-5b446d88c5-pvxdj\" (UID: \"2e8a774f-7876-4f56-94e6-c559397129c1\") " pod="cert-manager/cert-manager-5b446d88c5-pvxdj" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.363932 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nss7r\" (UniqueName: \"kubernetes.io/projected/9f1cba27-3a94-461f-bee4-1dde7d31f9a6-kube-api-access-nss7r\") pod \"cert-manager-webhook-5655c58dd6-ptpz9\" (UID: \"9f1cba27-3a94-461f-bee4-1dde7d31f9a6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-ptpz9" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.383073 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ncn2\" (UniqueName: \"kubernetes.io/projected/2e8a774f-7876-4f56-94e6-c559397129c1-kube-api-access-7ncn2\") pod \"cert-manager-5b446d88c5-pvxdj\" (UID: \"2e8a774f-7876-4f56-94e6-c559397129c1\") " pod="cert-manager/cert-manager-5b446d88c5-pvxdj" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.384534 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nss7r\" (UniqueName: \"kubernetes.io/projected/9f1cba27-3a94-461f-bee4-1dde7d31f9a6-kube-api-access-nss7r\") pod \"cert-manager-webhook-5655c58dd6-ptpz9\" (UID: \"9f1cba27-3a94-461f-bee4-1dde7d31f9a6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-ptpz9" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.415533 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-6dclx" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.454064 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-ptpz9" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.461793 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-pvxdj" Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.709346 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-6dclx"] Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.727562 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.837419 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-pvxdj"] Nov 27 10:30:45 crc kubenswrapper[4806]: W1127 10:30:45.845346 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e8a774f_7876_4f56_94e6_c559397129c1.slice/crio-b3ac094041695e5eabf5e79f6011d7b1c74e515fc9ee747373f332a367f979fe WatchSource:0}: Error finding container b3ac094041695e5eabf5e79f6011d7b1c74e515fc9ee747373f332a367f979fe: Status 404 returned error can't find the container with id b3ac094041695e5eabf5e79f6011d7b1c74e515fc9ee747373f332a367f979fe Nov 27 10:30:45 crc kubenswrapper[4806]: I1127 10:30:45.979851 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-ptpz9"] Nov 27 10:30:45 crc kubenswrapper[4806]: W1127 10:30:45.986659 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f1cba27_3a94_461f_bee4_1dde7d31f9a6.slice/crio-80cd8c317e617385468227674601c1c99d482d788fed887c7527c3ee553561d6 WatchSource:0}: Error finding container 80cd8c317e617385468227674601c1c99d482d788fed887c7527c3ee553561d6: Status 404 returned error can't find the container with id 80cd8c317e617385468227674601c1c99d482d788fed887c7527c3ee553561d6 Nov 27 10:30:46 crc kubenswrapper[4806]: I1127 10:30:46.474225 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-6dclx" event={"ID":"42836747-f4cd-43a2-a690-b17989a8d2c1","Type":"ContainerStarted","Data":"0cbde4f614da6abefc63d6375c84cc636fa04ec5dc8e1224f5c54b1e43c6c72d"} Nov 27 10:30:46 crc kubenswrapper[4806]: I1127 10:30:46.475711 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-pvxdj" event={"ID":"2e8a774f-7876-4f56-94e6-c559397129c1","Type":"ContainerStarted","Data":"b3ac094041695e5eabf5e79f6011d7b1c74e515fc9ee747373f332a367f979fe"} Nov 27 10:30:46 crc kubenswrapper[4806]: I1127 10:30:46.477360 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-ptpz9" event={"ID":"9f1cba27-3a94-461f-bee4-1dde7d31f9a6","Type":"ContainerStarted","Data":"80cd8c317e617385468227674601c1c99d482d788fed887c7527c3ee553561d6"} Nov 27 10:30:48 crc kubenswrapper[4806]: I1127 10:30:48.492254 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-6dclx" event={"ID":"42836747-f4cd-43a2-a690-b17989a8d2c1","Type":"ContainerStarted","Data":"c60484863ce7b546d6bf614b6eb6bdc04ecaaecd44f7417dedceb915780df0e0"} Nov 27 10:30:48 crc kubenswrapper[4806]: I1127 10:30:48.516102 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-6dclx" podStartSLOduration=1.3463338519999999 podStartE2EDuration="3.516083613s" podCreationTimestamp="2025-11-27 10:30:45 +0000 UTC" firstStartedPulling="2025-11-27 10:30:45.726453328 +0000 UTC m=+550.313044092" lastFinishedPulling="2025-11-27 10:30:47.896203079 +0000 UTC m=+552.482793853" observedRunningTime="2025-11-27 10:30:48.514781476 +0000 UTC m=+553.101372240" watchObservedRunningTime="2025-11-27 10:30:48.516083613 +0000 UTC m=+553.102674377" Nov 27 10:30:50 crc kubenswrapper[4806]: I1127 10:30:50.503738 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-pvxdj" event={"ID":"2e8a774f-7876-4f56-94e6-c559397129c1","Type":"ContainerStarted","Data":"459d58a09fd15c45f29a1c193b96b6c14ee2ffc28d454154ec1f448436d4e5ca"} Nov 27 10:30:50 crc kubenswrapper[4806]: I1127 10:30:50.506373 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-ptpz9" event={"ID":"9f1cba27-3a94-461f-bee4-1dde7d31f9a6","Type":"ContainerStarted","Data":"ec8318e483905d412ea4be77b0b5da07bfb4bda45725ceba2ff6fd79ee668366"} Nov 27 10:30:50 crc kubenswrapper[4806]: I1127 10:30:50.506550 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-ptpz9" Nov 27 10:30:50 crc kubenswrapper[4806]: I1127 10:30:50.525190 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-pvxdj" podStartSLOduration=1.5777533670000001 podStartE2EDuration="5.525166859s" podCreationTimestamp="2025-11-27 10:30:45 +0000 UTC" firstStartedPulling="2025-11-27 10:30:45.848155962 +0000 UTC m=+550.434746726" lastFinishedPulling="2025-11-27 10:30:49.795569454 +0000 UTC m=+554.382160218" observedRunningTime="2025-11-27 10:30:50.523279645 +0000 UTC m=+555.109870409" watchObservedRunningTime="2025-11-27 10:30:50.525166859 +0000 UTC m=+555.111757623" Nov 27 10:30:50 crc kubenswrapper[4806]: I1127 10:30:50.547727 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-ptpz9" podStartSLOduration=1.74077002 podStartE2EDuration="5.547702483s" podCreationTimestamp="2025-11-27 10:30:45 +0000 UTC" firstStartedPulling="2025-11-27 10:30:45.990056942 +0000 UTC m=+550.576647706" lastFinishedPulling="2025-11-27 10:30:49.796989415 +0000 UTC m=+554.383580169" observedRunningTime="2025-11-27 10:30:50.54622449 +0000 UTC m=+555.132815274" watchObservedRunningTime="2025-11-27 10:30:50.547702483 +0000 UTC m=+555.134293247" Nov 27 10:30:55 crc kubenswrapper[4806]: I1127 10:30:55.458379 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-ptpz9" Nov 27 10:30:55 crc kubenswrapper[4806]: I1127 10:30:55.790668 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rsgmm"] Nov 27 10:30:55 crc kubenswrapper[4806]: I1127 10:30:55.791210 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20" gracePeriod=30 Nov 27 10:30:55 crc kubenswrapper[4806]: I1127 10:30:55.791322 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="kube-rbac-proxy-node" containerID="cri-o://19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73" gracePeriod=30 Nov 27 10:30:55 crc kubenswrapper[4806]: I1127 10:30:55.791399 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="northd" containerID="cri-o://761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc" gracePeriod=30 Nov 27 10:30:55 crc kubenswrapper[4806]: I1127 10:30:55.791516 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovn-acl-logging" containerID="cri-o://1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7" gracePeriod=30 Nov 27 10:30:55 crc kubenswrapper[4806]: I1127 10:30:55.791573 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="sbdb" containerID="cri-o://64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef" gracePeriod=30 Nov 27 10:30:55 crc kubenswrapper[4806]: I1127 10:30:55.794451 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="nbdb" containerID="cri-o://1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744" gracePeriod=30 Nov 27 10:30:55 crc kubenswrapper[4806]: I1127 10:30:55.794541 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovn-controller" containerID="cri-o://f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b" gracePeriod=30 Nov 27 10:30:55 crc kubenswrapper[4806]: I1127 10:30:55.865543 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" containerID="cri-o://e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b" gracePeriod=30 Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.209265 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/3.log" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.213109 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovn-acl-logging/0.log" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.213680 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovn-controller/0.log" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.214276 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.261486 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovnkube-config\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.261844 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-etc-openvswitch\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.261936 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-run-netns\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262182 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-systemd-units\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262221 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-cni-netd\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262287 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-systemd\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262328 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-env-overrides\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262364 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovn-node-metrics-cert\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262408 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-var-lib-openvswitch\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262449 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262490 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-run-ovn-kubernetes\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262533 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czmdn\" (UniqueName: \"kubernetes.io/projected/e8ed6bbc-965a-412c-a3ee-4763035aa34e-kube-api-access-czmdn\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262594 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-slash\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262632 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-ovn\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262683 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-log-socket\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262734 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovnkube-script-lib\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262784 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-cni-bin\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262824 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-node-log\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262861 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-openvswitch\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.262899 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-kubelet\") pod \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\" (UID: \"e8ed6bbc-965a-412c-a3ee-4763035aa34e\") " Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.263336 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.263526 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.263592 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.263645 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-slash" (OuterVolumeSpecName: "host-slash") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.263685 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.263686 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.264158 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.264164 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.264192 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.264222 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.264225 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.264262 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-log-socket" (OuterVolumeSpecName: "log-socket") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.264283 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-node-log" (OuterVolumeSpecName: "node-log") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.264311 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.264340 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.264633 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.266277 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.276194 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.276607 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ed6bbc-965a-412c-a3ee-4763035aa34e-kube-api-access-czmdn" (OuterVolumeSpecName: "kube-api-access-czmdn") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "kube-api-access-czmdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.287318 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4nc2r"] Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.287661 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovn-acl-logging" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.287687 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovn-acl-logging" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.287703 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.287725 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.287736 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovn-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.287745 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovn-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.287758 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.287771 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.287782 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="nbdb" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.287791 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="nbdb" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.287807 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="kube-rbac-proxy-ovn-metrics" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.287817 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="kube-rbac-proxy-ovn-metrics" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.287828 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="northd" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.287837 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="northd" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.287849 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="kubecfg-setup" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.287857 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="kubecfg-setup" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.287871 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.287879 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.287892 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.287900 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.287914 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="kube-rbac-proxy-node" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.287922 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="kube-rbac-proxy-node" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.287934 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="sbdb" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.287943 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="sbdb" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.288064 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.288078 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.288090 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.288100 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovn-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.288111 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="kube-rbac-proxy-ovn-metrics" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.288126 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.288142 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="sbdb" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.288160 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovn-acl-logging" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.288172 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="northd" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.288183 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="kube-rbac-proxy-node" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.288198 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="nbdb" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.288405 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.288417 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.288556 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerName="ovnkube-controller" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.290857 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "e8ed6bbc-965a-412c-a3ee-4763035aa34e" (UID: "e8ed6bbc-965a-412c-a3ee-4763035aa34e"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.291124 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364039 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-run-ovn-kubernetes\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364101 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ba802e41-04f3-4815-bce5-9633457bae75-env-overrides\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364122 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-run-openvswitch\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364144 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ba802e41-04f3-4815-bce5-9633457bae75-ovnkube-script-lib\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364296 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fclhm\" (UniqueName: \"kubernetes.io/projected/ba802e41-04f3-4815-bce5-9633457bae75-kube-api-access-fclhm\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364333 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-slash\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364364 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ba802e41-04f3-4815-bce5-9633457bae75-ovn-node-metrics-cert\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364380 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-etc-openvswitch\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364399 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-systemd-units\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364414 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-node-log\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364437 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-run-systemd\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364458 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-run-netns\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364474 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ba802e41-04f3-4815-bce5-9633457bae75-ovnkube-config\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364493 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364508 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-kubelet\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364529 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-var-lib-openvswitch\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364548 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-log-socket\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364567 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-run-ovn\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364587 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-cni-bin\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364604 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-cni-netd\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364643 4806 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-slash\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364656 4806 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364685 4806 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-log-socket\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364696 4806 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364705 4806 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364713 4806 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-node-log\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364722 4806 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364731 4806 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364740 4806 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364749 4806 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364758 4806 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364768 4806 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364776 4806 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364790 4806 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364798 4806 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e8ed6bbc-965a-412c-a3ee-4763035aa34e-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364807 4806 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e8ed6bbc-965a-412c-a3ee-4763035aa34e-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364817 4806 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364827 4806 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364836 4806 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e8ed6bbc-965a-412c-a3ee-4763035aa34e-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.364845 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czmdn\" (UniqueName: \"kubernetes.io/projected/e8ed6bbc-965a-412c-a3ee-4763035aa34e-kube-api-access-czmdn\") on node \"crc\" DevicePath \"\"" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465396 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-systemd-units\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465438 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-node-log\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465461 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-run-systemd\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465479 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-run-netns\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465498 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ba802e41-04f3-4815-bce5-9633457bae75-ovnkube-config\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465520 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465540 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-kubelet\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465555 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-var-lib-openvswitch\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465575 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-log-socket\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465592 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-run-ovn\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465610 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-cni-bin\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465624 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-cni-netd\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465651 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-run-ovn-kubernetes\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465665 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ba802e41-04f3-4815-bce5-9633457bae75-env-overrides\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465680 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-run-openvswitch\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465697 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fclhm\" (UniqueName: \"kubernetes.io/projected/ba802e41-04f3-4815-bce5-9633457bae75-kube-api-access-fclhm\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465713 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ba802e41-04f3-4815-bce5-9633457bae75-ovnkube-script-lib\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465740 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-slash\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465759 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ba802e41-04f3-4815-bce5-9633457bae75-ovn-node-metrics-cert\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465817 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-etc-openvswitch\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465856 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-var-lib-openvswitch\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465898 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-etc-openvswitch\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465953 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-log-socket\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465953 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-systemd-units\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465994 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-cni-bin\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.465997 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-run-ovn\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.466035 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-cni-netd\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.466044 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-node-log\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.466075 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-run-ovn-kubernetes\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.466085 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-run-systemd\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.466127 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-run-netns\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.466823 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ba802e41-04f3-4815-bce5-9633457bae75-env-overrides\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.466870 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-run-openvswitch\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.467138 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ba802e41-04f3-4815-bce5-9633457bae75-ovnkube-config\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.467274 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.467297 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-kubelet\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.467338 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ba802e41-04f3-4815-bce5-9633457bae75-host-slash\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.468264 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ba802e41-04f3-4815-bce5-9633457bae75-ovnkube-script-lib\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.473635 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ba802e41-04f3-4815-bce5-9633457bae75-ovn-node-metrics-cert\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.494424 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fclhm\" (UniqueName: \"kubernetes.io/projected/ba802e41-04f3-4815-bce5-9633457bae75-kube-api-access-fclhm\") pod \"ovnkube-node-4nc2r\" (UID: \"ba802e41-04f3-4815-bce5-9633457bae75\") " pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.567075 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5cx6g_39930cf4-5f3c-42cc-9384-d366bb1d6a78/kube-multus/2.log" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.567748 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5cx6g_39930cf4-5f3c-42cc-9384-d366bb1d6a78/kube-multus/1.log" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.567843 4806 generic.go:334] "Generic (PLEG): container finished" podID="39930cf4-5f3c-42cc-9384-d366bb1d6a78" containerID="6e8529945bfb913b81082452411a2bfcf1b07dde94eaa1f2edce41149edb767e" exitCode=2 Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.567951 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5cx6g" event={"ID":"39930cf4-5f3c-42cc-9384-d366bb1d6a78","Type":"ContainerDied","Data":"6e8529945bfb913b81082452411a2bfcf1b07dde94eaa1f2edce41149edb767e"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.568043 4806 scope.go:117] "RemoveContainer" containerID="04e5f8cbf4179151c6868e9168364cc8fa754b47829422ce6a64fa25edf7a949" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.568509 4806 scope.go:117] "RemoveContainer" containerID="6e8529945bfb913b81082452411a2bfcf1b07dde94eaa1f2edce41149edb767e" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.570179 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-5cx6g_openshift-multus(39930cf4-5f3c-42cc-9384-d366bb1d6a78)\"" pod="openshift-multus/multus-5cx6g" podUID="39930cf4-5f3c-42cc-9384-d366bb1d6a78" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.571960 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovnkube-controller/3.log" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.575451 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovn-acl-logging/0.log" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577282 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rsgmm_e8ed6bbc-965a-412c-a3ee-4763035aa34e/ovn-controller/0.log" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577645 4806 generic.go:334] "Generic (PLEG): container finished" podID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerID="e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b" exitCode=0 Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577684 4806 generic.go:334] "Generic (PLEG): container finished" podID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerID="64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef" exitCode=0 Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577696 4806 generic.go:334] "Generic (PLEG): container finished" podID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerID="1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744" exitCode=0 Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577709 4806 generic.go:334] "Generic (PLEG): container finished" podID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerID="761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc" exitCode=0 Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577717 4806 generic.go:334] "Generic (PLEG): container finished" podID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerID="c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20" exitCode=0 Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577724 4806 generic.go:334] "Generic (PLEG): container finished" podID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerID="19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73" exitCode=0 Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577732 4806 generic.go:334] "Generic (PLEG): container finished" podID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerID="1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7" exitCode=143 Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577740 4806 generic.go:334] "Generic (PLEG): container finished" podID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" containerID="f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b" exitCode=143 Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577765 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577805 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577816 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577827 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577860 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577872 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577885 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577900 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577907 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577915 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577921 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577927 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577933 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577939 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577946 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577953 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577962 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577972 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577980 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577986 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577992 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577999 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578004 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578010 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578017 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578023 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578029 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578038 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578048 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578055 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578061 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578067 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578073 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578079 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578085 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578092 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578098 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578104 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578112 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" event={"ID":"e8ed6bbc-965a-412c-a3ee-4763035aa34e","Type":"ContainerDied","Data":"89943e205c0bc21f519cd22041759a0a025bfee3dfdc5e836e5939bf1b2b2232"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578120 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578129 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578135 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578141 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578148 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578154 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578160 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578166 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578172 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.578177 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961"} Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.577874 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rsgmm" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.607816 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.615387 4806 scope.go:117] "RemoveContainer" containerID="e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.634271 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rsgmm"] Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.637630 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rsgmm"] Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.649197 4806 scope.go:117] "RemoveContainer" containerID="faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.667207 4806 scope.go:117] "RemoveContainer" containerID="64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.688984 4806 scope.go:117] "RemoveContainer" containerID="1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.703500 4806 scope.go:117] "RemoveContainer" containerID="761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.718709 4806 scope.go:117] "RemoveContainer" containerID="c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.751827 4806 scope.go:117] "RemoveContainer" containerID="19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.789300 4806 scope.go:117] "RemoveContainer" containerID="1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.806431 4806 scope.go:117] "RemoveContainer" containerID="f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.877462 4806 scope.go:117] "RemoveContainer" containerID="2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.898276 4806 scope.go:117] "RemoveContainer" containerID="e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.907634 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b\": container with ID starting with e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b not found: ID does not exist" containerID="e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.907706 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b"} err="failed to get container status \"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b\": rpc error: code = NotFound desc = could not find container \"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b\": container with ID starting with e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.907751 4806 scope.go:117] "RemoveContainer" containerID="faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.908398 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\": container with ID starting with faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084 not found: ID does not exist" containerID="faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.908427 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084"} err="failed to get container status \"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\": rpc error: code = NotFound desc = could not find container \"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\": container with ID starting with faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.908445 4806 scope.go:117] "RemoveContainer" containerID="64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.909538 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\": container with ID starting with 64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef not found: ID does not exist" containerID="64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.909590 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef"} err="failed to get container status \"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\": rpc error: code = NotFound desc = could not find container \"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\": container with ID starting with 64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.909639 4806 scope.go:117] "RemoveContainer" containerID="1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.910057 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\": container with ID starting with 1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744 not found: ID does not exist" containerID="1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.910115 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744"} err="failed to get container status \"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\": rpc error: code = NotFound desc = could not find container \"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\": container with ID starting with 1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.910153 4806 scope.go:117] "RemoveContainer" containerID="761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.910579 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\": container with ID starting with 761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc not found: ID does not exist" containerID="761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.910611 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc"} err="failed to get container status \"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\": rpc error: code = NotFound desc = could not find container \"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\": container with ID starting with 761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.910626 4806 scope.go:117] "RemoveContainer" containerID="c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.911107 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\": container with ID starting with c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20 not found: ID does not exist" containerID="c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.911141 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20"} err="failed to get container status \"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\": rpc error: code = NotFound desc = could not find container \"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\": container with ID starting with c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.911160 4806 scope.go:117] "RemoveContainer" containerID="19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.911495 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\": container with ID starting with 19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73 not found: ID does not exist" containerID="19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.911540 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73"} err="failed to get container status \"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\": rpc error: code = NotFound desc = could not find container \"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\": container with ID starting with 19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.911727 4806 scope.go:117] "RemoveContainer" containerID="1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.912302 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\": container with ID starting with 1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7 not found: ID does not exist" containerID="1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.912594 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7"} err="failed to get container status \"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\": rpc error: code = NotFound desc = could not find container \"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\": container with ID starting with 1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.912609 4806 scope.go:117] "RemoveContainer" containerID="f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.913364 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\": container with ID starting with f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b not found: ID does not exist" containerID="f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.913444 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b"} err="failed to get container status \"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\": rpc error: code = NotFound desc = could not find container \"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\": container with ID starting with f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.913501 4806 scope.go:117] "RemoveContainer" containerID="2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961" Nov 27 10:30:56 crc kubenswrapper[4806]: E1127 10:30:56.914100 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\": container with ID starting with 2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961 not found: ID does not exist" containerID="2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.914126 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961"} err="failed to get container status \"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\": rpc error: code = NotFound desc = could not find container \"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\": container with ID starting with 2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.914143 4806 scope.go:117] "RemoveContainer" containerID="e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.915055 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b"} err="failed to get container status \"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b\": rpc error: code = NotFound desc = could not find container \"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b\": container with ID starting with e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.915073 4806 scope.go:117] "RemoveContainer" containerID="faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.915791 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084"} err="failed to get container status \"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\": rpc error: code = NotFound desc = could not find container \"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\": container with ID starting with faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.915851 4806 scope.go:117] "RemoveContainer" containerID="64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.916316 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef"} err="failed to get container status \"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\": rpc error: code = NotFound desc = could not find container \"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\": container with ID starting with 64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.916347 4806 scope.go:117] "RemoveContainer" containerID="1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.916627 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744"} err="failed to get container status \"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\": rpc error: code = NotFound desc = could not find container \"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\": container with ID starting with 1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.916651 4806 scope.go:117] "RemoveContainer" containerID="761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.916945 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc"} err="failed to get container status \"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\": rpc error: code = NotFound desc = could not find container \"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\": container with ID starting with 761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.916966 4806 scope.go:117] "RemoveContainer" containerID="c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.917326 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20"} err="failed to get container status \"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\": rpc error: code = NotFound desc = could not find container \"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\": container with ID starting with c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.917357 4806 scope.go:117] "RemoveContainer" containerID="19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.919628 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73"} err="failed to get container status \"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\": rpc error: code = NotFound desc = could not find container \"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\": container with ID starting with 19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.919683 4806 scope.go:117] "RemoveContainer" containerID="1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.920456 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7"} err="failed to get container status \"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\": rpc error: code = NotFound desc = could not find container \"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\": container with ID starting with 1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.920543 4806 scope.go:117] "RemoveContainer" containerID="f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.921346 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b"} err="failed to get container status \"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\": rpc error: code = NotFound desc = could not find container \"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\": container with ID starting with f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.921374 4806 scope.go:117] "RemoveContainer" containerID="2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.922100 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961"} err="failed to get container status \"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\": rpc error: code = NotFound desc = could not find container \"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\": container with ID starting with 2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.922137 4806 scope.go:117] "RemoveContainer" containerID="e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.923062 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b"} err="failed to get container status \"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b\": rpc error: code = NotFound desc = could not find container \"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b\": container with ID starting with e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.923117 4806 scope.go:117] "RemoveContainer" containerID="faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.923521 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084"} err="failed to get container status \"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\": rpc error: code = NotFound desc = could not find container \"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\": container with ID starting with faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.923554 4806 scope.go:117] "RemoveContainer" containerID="64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.924025 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef"} err="failed to get container status \"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\": rpc error: code = NotFound desc = could not find container \"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\": container with ID starting with 64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.924064 4806 scope.go:117] "RemoveContainer" containerID="1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.924647 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744"} err="failed to get container status \"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\": rpc error: code = NotFound desc = could not find container \"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\": container with ID starting with 1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.924674 4806 scope.go:117] "RemoveContainer" containerID="761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.925101 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc"} err="failed to get container status \"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\": rpc error: code = NotFound desc = could not find container \"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\": container with ID starting with 761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.925132 4806 scope.go:117] "RemoveContainer" containerID="c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.925506 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20"} err="failed to get container status \"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\": rpc error: code = NotFound desc = could not find container \"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\": container with ID starting with c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.925544 4806 scope.go:117] "RemoveContainer" containerID="19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.927170 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73"} err="failed to get container status \"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\": rpc error: code = NotFound desc = could not find container \"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\": container with ID starting with 19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.927197 4806 scope.go:117] "RemoveContainer" containerID="1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.927619 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7"} err="failed to get container status \"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\": rpc error: code = NotFound desc = could not find container \"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\": container with ID starting with 1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.927651 4806 scope.go:117] "RemoveContainer" containerID="f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.928079 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b"} err="failed to get container status \"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\": rpc error: code = NotFound desc = could not find container \"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\": container with ID starting with f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.928119 4806 scope.go:117] "RemoveContainer" containerID="2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.928565 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961"} err="failed to get container status \"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\": rpc error: code = NotFound desc = could not find container \"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\": container with ID starting with 2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.928603 4806 scope.go:117] "RemoveContainer" containerID="e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.929068 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b"} err="failed to get container status \"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b\": rpc error: code = NotFound desc = could not find container \"e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b\": container with ID starting with e9efd9fb737857404a4f6872b255431dd86af3de339d9f1a861c2143d75e4a0b not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.929090 4806 scope.go:117] "RemoveContainer" containerID="faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.929697 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084"} err="failed to get container status \"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\": rpc error: code = NotFound desc = could not find container \"faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084\": container with ID starting with faf98a9e4015bee3ee18298784bcf3c634d58c607a9b86d7abd52559b088d084 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.929794 4806 scope.go:117] "RemoveContainer" containerID="64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.933581 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef"} err="failed to get container status \"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\": rpc error: code = NotFound desc = could not find container \"64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef\": container with ID starting with 64fdb18865aada89a9d9bc3c9b3dd0438e410e064684514d45bd803113bcebef not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.933607 4806 scope.go:117] "RemoveContainer" containerID="1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.933917 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744"} err="failed to get container status \"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\": rpc error: code = NotFound desc = could not find container \"1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744\": container with ID starting with 1e73b3e694613d07e3a2c5024fc028de283031a6bd661273567d4e1162258744 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.933951 4806 scope.go:117] "RemoveContainer" containerID="761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.934210 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc"} err="failed to get container status \"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\": rpc error: code = NotFound desc = could not find container \"761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc\": container with ID starting with 761415deefb9c924b679b624f69da4ff6a2e066a3f9ead905f40991d476deadc not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.934225 4806 scope.go:117] "RemoveContainer" containerID="c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.934677 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20"} err="failed to get container status \"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\": rpc error: code = NotFound desc = could not find container \"c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20\": container with ID starting with c21d8e1df995e0531e02ab50e1b429658377ed2c0c47eb9122b8df1a0d488e20 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.934714 4806 scope.go:117] "RemoveContainer" containerID="19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.936500 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73"} err="failed to get container status \"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\": rpc error: code = NotFound desc = could not find container \"19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73\": container with ID starting with 19fcd98b0969b2b0960af683688c0a23ac43f7da887ac71ccef44121b9d0fc73 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.936531 4806 scope.go:117] "RemoveContainer" containerID="1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.938376 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7"} err="failed to get container status \"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\": rpc error: code = NotFound desc = could not find container \"1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7\": container with ID starting with 1894a842b08a4a3a0fbfc876f69b1ecec49f84cd247021340609a136fabf67e7 not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.938409 4806 scope.go:117] "RemoveContainer" containerID="f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.940430 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b"} err="failed to get container status \"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\": rpc error: code = NotFound desc = could not find container \"f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b\": container with ID starting with f083e96af48817e9157b41db76e2d2b7f8b372f07fd2f17013c11823493e578b not found: ID does not exist" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.940479 4806 scope.go:117] "RemoveContainer" containerID="2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961" Nov 27 10:30:56 crc kubenswrapper[4806]: I1127 10:30:56.941004 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961"} err="failed to get container status \"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\": rpc error: code = NotFound desc = could not find container \"2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961\": container with ID starting with 2f0ac78992e6d574ccea3937dede46c3a5138bac6a1d054d01fd0df0770e0961 not found: ID does not exist" Nov 27 10:30:57 crc kubenswrapper[4806]: I1127 10:30:57.599551 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5cx6g_39930cf4-5f3c-42cc-9384-d366bb1d6a78/kube-multus/2.log" Nov 27 10:30:57 crc kubenswrapper[4806]: I1127 10:30:57.603843 4806 generic.go:334] "Generic (PLEG): container finished" podID="ba802e41-04f3-4815-bce5-9633457bae75" containerID="cdf70bd738593610661cf36c046cfa99f0f53d46eb1cb621cfa0038f5acb7cfe" exitCode=0 Nov 27 10:30:57 crc kubenswrapper[4806]: I1127 10:30:57.603884 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" event={"ID":"ba802e41-04f3-4815-bce5-9633457bae75","Type":"ContainerDied","Data":"cdf70bd738593610661cf36c046cfa99f0f53d46eb1cb621cfa0038f5acb7cfe"} Nov 27 10:30:57 crc kubenswrapper[4806]: I1127 10:30:57.603909 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" event={"ID":"ba802e41-04f3-4815-bce5-9633457bae75","Type":"ContainerStarted","Data":"e2a5f0279d4083adc477eca4c055aa6c558b098e6fa0b00c7d12378dc1c79131"} Nov 27 10:30:58 crc kubenswrapper[4806]: I1127 10:30:58.125361 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8ed6bbc-965a-412c-a3ee-4763035aa34e" path="/var/lib/kubelet/pods/e8ed6bbc-965a-412c-a3ee-4763035aa34e/volumes" Nov 27 10:30:58 crc kubenswrapper[4806]: I1127 10:30:58.616492 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" event={"ID":"ba802e41-04f3-4815-bce5-9633457bae75","Type":"ContainerStarted","Data":"04c5c261a6d99f4ecf869471506d5636dc9d6273312e384f5657a5c23ff3989f"} Nov 27 10:30:58 crc kubenswrapper[4806]: I1127 10:30:58.616574 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" event={"ID":"ba802e41-04f3-4815-bce5-9633457bae75","Type":"ContainerStarted","Data":"e2708016d44624d59cda294bf2170b1ce5b0e1fe3de18367b4854fd23f80d14a"} Nov 27 10:30:58 crc kubenswrapper[4806]: I1127 10:30:58.616596 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" event={"ID":"ba802e41-04f3-4815-bce5-9633457bae75","Type":"ContainerStarted","Data":"9b30c4cbb1b7db8d6443a0df5d8e6f65fe16467d9677d11974692afb5d885566"} Nov 27 10:30:58 crc kubenswrapper[4806]: I1127 10:30:58.616612 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" event={"ID":"ba802e41-04f3-4815-bce5-9633457bae75","Type":"ContainerStarted","Data":"c0e608b311393e1deffaedd8983bb039b17f3c5e37b12a02cb75649c605a9c18"} Nov 27 10:30:58 crc kubenswrapper[4806]: I1127 10:30:58.616625 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" event={"ID":"ba802e41-04f3-4815-bce5-9633457bae75","Type":"ContainerStarted","Data":"2ff44c83ce63164009e6ff535706cf5da4198e45fb6833a9a595c9ccce5734ac"} Nov 27 10:30:58 crc kubenswrapper[4806]: I1127 10:30:58.616639 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" event={"ID":"ba802e41-04f3-4815-bce5-9633457bae75","Type":"ContainerStarted","Data":"affcb7f513d3228edd6947689de3a927a5f3cf3572ad018170d497b3c07c2d0c"} Nov 27 10:31:01 crc kubenswrapper[4806]: I1127 10:31:01.646467 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" event={"ID":"ba802e41-04f3-4815-bce5-9633457bae75","Type":"ContainerStarted","Data":"0985eb35c3c43b86f17046887b8f2bab49dd25a48cdaff146d368ce3d73dc60e"} Nov 27 10:31:03 crc kubenswrapper[4806]: I1127 10:31:03.663167 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" event={"ID":"ba802e41-04f3-4815-bce5-9633457bae75","Type":"ContainerStarted","Data":"665cb1622a04e9a97c12d25fa6415e25f6580b8912cad0cd976bedbcbe3b1699"} Nov 27 10:31:03 crc kubenswrapper[4806]: I1127 10:31:03.663886 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:31:03 crc kubenswrapper[4806]: I1127 10:31:03.663900 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:31:03 crc kubenswrapper[4806]: I1127 10:31:03.703432 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:31:03 crc kubenswrapper[4806]: I1127 10:31:03.709882 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" podStartSLOduration=7.709860823 podStartE2EDuration="7.709860823s" podCreationTimestamp="2025-11-27 10:30:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:31:03.704771817 +0000 UTC m=+568.291362581" watchObservedRunningTime="2025-11-27 10:31:03.709860823 +0000 UTC m=+568.296451587" Nov 27 10:31:04 crc kubenswrapper[4806]: I1127 10:31:04.672278 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:31:04 crc kubenswrapper[4806]: I1127 10:31:04.713828 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:31:11 crc kubenswrapper[4806]: I1127 10:31:11.117366 4806 scope.go:117] "RemoveContainer" containerID="6e8529945bfb913b81082452411a2bfcf1b07dde94eaa1f2edce41149edb767e" Nov 27 10:31:11 crc kubenswrapper[4806]: E1127 10:31:11.117837 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-5cx6g_openshift-multus(39930cf4-5f3c-42cc-9384-d366bb1d6a78)\"" pod="openshift-multus/multus-5cx6g" podUID="39930cf4-5f3c-42cc-9384-d366bb1d6a78" Nov 27 10:31:14 crc kubenswrapper[4806]: I1127 10:31:14.450739 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:31:14 crc kubenswrapper[4806]: I1127 10:31:14.451090 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:31:14 crc kubenswrapper[4806]: I1127 10:31:14.451177 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:31:14 crc kubenswrapper[4806]: I1127 10:31:14.452189 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"27aa6f365485f368ff18494a5978d93bb6e961422bea73da5d7f2d1830e740c0"} pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 10:31:14 crc kubenswrapper[4806]: I1127 10:31:14.452329 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" containerID="cri-o://27aa6f365485f368ff18494a5978d93bb6e961422bea73da5d7f2d1830e740c0" gracePeriod=600 Nov 27 10:31:14 crc kubenswrapper[4806]: I1127 10:31:14.730985 4806 generic.go:334] "Generic (PLEG): container finished" podID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerID="27aa6f365485f368ff18494a5978d93bb6e961422bea73da5d7f2d1830e740c0" exitCode=0 Nov 27 10:31:14 crc kubenswrapper[4806]: I1127 10:31:14.731147 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerDied","Data":"27aa6f365485f368ff18494a5978d93bb6e961422bea73da5d7f2d1830e740c0"} Nov 27 10:31:14 crc kubenswrapper[4806]: I1127 10:31:14.731874 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"5e15c67a8ce5b7fd1f5bca346e1724cf329b9d00aa70feed8824c46529a879e6"} Nov 27 10:31:14 crc kubenswrapper[4806]: I1127 10:31:14.731907 4806 scope.go:117] "RemoveContainer" containerID="0f2788d6ded82e196b1f288174c14861105e5b2e48bf181245fe8449fb3b1acb" Nov 27 10:31:22 crc kubenswrapper[4806]: I1127 10:31:22.117489 4806 scope.go:117] "RemoveContainer" containerID="6e8529945bfb913b81082452411a2bfcf1b07dde94eaa1f2edce41149edb767e" Nov 27 10:31:22 crc kubenswrapper[4806]: I1127 10:31:22.787729 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5cx6g_39930cf4-5f3c-42cc-9384-d366bb1d6a78/kube-multus/2.log" Nov 27 10:31:22 crc kubenswrapper[4806]: I1127 10:31:22.788721 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5cx6g" event={"ID":"39930cf4-5f3c-42cc-9384-d366bb1d6a78","Type":"ContainerStarted","Data":"5adf10641c85c9fe6c500b7f78d235c13f978fbaa11ccbac408591b07531f9db"} Nov 27 10:31:26 crc kubenswrapper[4806]: I1127 10:31:26.640524 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4nc2r" Nov 27 10:31:36 crc kubenswrapper[4806]: I1127 10:31:36.702980 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2"] Nov 27 10:31:36 crc kubenswrapper[4806]: I1127 10:31:36.705138 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" Nov 27 10:31:36 crc kubenswrapper[4806]: I1127 10:31:36.707321 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 27 10:31:36 crc kubenswrapper[4806]: I1127 10:31:36.712945 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2"] Nov 27 10:31:36 crc kubenswrapper[4806]: I1127 10:31:36.763653 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dd8e785-4b1d-406d-bb97-391b74415228-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2\" (UID: \"4dd8e785-4b1d-406d-bb97-391b74415228\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" Nov 27 10:31:36 crc kubenswrapper[4806]: I1127 10:31:36.763726 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dd8e785-4b1d-406d-bb97-391b74415228-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2\" (UID: \"4dd8e785-4b1d-406d-bb97-391b74415228\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" Nov 27 10:31:36 crc kubenswrapper[4806]: I1127 10:31:36.763782 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v95f\" (UniqueName: \"kubernetes.io/projected/4dd8e785-4b1d-406d-bb97-391b74415228-kube-api-access-6v95f\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2\" (UID: \"4dd8e785-4b1d-406d-bb97-391b74415228\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" Nov 27 10:31:36 crc kubenswrapper[4806]: I1127 10:31:36.865711 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v95f\" (UniqueName: \"kubernetes.io/projected/4dd8e785-4b1d-406d-bb97-391b74415228-kube-api-access-6v95f\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2\" (UID: \"4dd8e785-4b1d-406d-bb97-391b74415228\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" Nov 27 10:31:36 crc kubenswrapper[4806]: I1127 10:31:36.866137 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dd8e785-4b1d-406d-bb97-391b74415228-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2\" (UID: \"4dd8e785-4b1d-406d-bb97-391b74415228\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" Nov 27 10:31:36 crc kubenswrapper[4806]: I1127 10:31:36.866415 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dd8e785-4b1d-406d-bb97-391b74415228-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2\" (UID: \"4dd8e785-4b1d-406d-bb97-391b74415228\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" Nov 27 10:31:36 crc kubenswrapper[4806]: I1127 10:31:36.866632 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dd8e785-4b1d-406d-bb97-391b74415228-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2\" (UID: \"4dd8e785-4b1d-406d-bb97-391b74415228\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" Nov 27 10:31:36 crc kubenswrapper[4806]: I1127 10:31:36.866973 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dd8e785-4b1d-406d-bb97-391b74415228-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2\" (UID: \"4dd8e785-4b1d-406d-bb97-391b74415228\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" Nov 27 10:31:36 crc kubenswrapper[4806]: I1127 10:31:36.892820 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v95f\" (UniqueName: \"kubernetes.io/projected/4dd8e785-4b1d-406d-bb97-391b74415228-kube-api-access-6v95f\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2\" (UID: \"4dd8e785-4b1d-406d-bb97-391b74415228\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" Nov 27 10:31:37 crc kubenswrapper[4806]: I1127 10:31:37.023933 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" Nov 27 10:31:37 crc kubenswrapper[4806]: I1127 10:31:37.288500 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2"] Nov 27 10:31:37 crc kubenswrapper[4806]: W1127 10:31:37.296217 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4dd8e785_4b1d_406d_bb97_391b74415228.slice/crio-156b2534d1998f17fd945f399f687c4337480572a77e92ebc892e0846fad45b5 WatchSource:0}: Error finding container 156b2534d1998f17fd945f399f687c4337480572a77e92ebc892e0846fad45b5: Status 404 returned error can't find the container with id 156b2534d1998f17fd945f399f687c4337480572a77e92ebc892e0846fad45b5 Nov 27 10:31:37 crc kubenswrapper[4806]: I1127 10:31:37.889464 4806 generic.go:334] "Generic (PLEG): container finished" podID="4dd8e785-4b1d-406d-bb97-391b74415228" containerID="df7b1db5b2a3d4267330ce9f6449ef0d20b252f224e5f7b7bca97ee2dab45fce" exitCode=0 Nov 27 10:31:37 crc kubenswrapper[4806]: I1127 10:31:37.890014 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" event={"ID":"4dd8e785-4b1d-406d-bb97-391b74415228","Type":"ContainerDied","Data":"df7b1db5b2a3d4267330ce9f6449ef0d20b252f224e5f7b7bca97ee2dab45fce"} Nov 27 10:31:37 crc kubenswrapper[4806]: I1127 10:31:37.890068 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" event={"ID":"4dd8e785-4b1d-406d-bb97-391b74415228","Type":"ContainerStarted","Data":"156b2534d1998f17fd945f399f687c4337480572a77e92ebc892e0846fad45b5"} Nov 27 10:31:39 crc kubenswrapper[4806]: I1127 10:31:39.904383 4806 generic.go:334] "Generic (PLEG): container finished" podID="4dd8e785-4b1d-406d-bb97-391b74415228" containerID="598af0f37fa00edefa6567128c9c541508587823578847cd59815849d9d82a5f" exitCode=0 Nov 27 10:31:39 crc kubenswrapper[4806]: I1127 10:31:39.904442 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" event={"ID":"4dd8e785-4b1d-406d-bb97-391b74415228","Type":"ContainerDied","Data":"598af0f37fa00edefa6567128c9c541508587823578847cd59815849d9d82a5f"} Nov 27 10:31:40 crc kubenswrapper[4806]: I1127 10:31:40.916579 4806 generic.go:334] "Generic (PLEG): container finished" podID="4dd8e785-4b1d-406d-bb97-391b74415228" containerID="73c4f2984deb70d461de1d3bdece10c88f5d3db0e19d9c9deac40014315ee1b2" exitCode=0 Nov 27 10:31:40 crc kubenswrapper[4806]: I1127 10:31:40.916744 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" event={"ID":"4dd8e785-4b1d-406d-bb97-391b74415228","Type":"ContainerDied","Data":"73c4f2984deb70d461de1d3bdece10c88f5d3db0e19d9c9deac40014315ee1b2"} Nov 27 10:31:42 crc kubenswrapper[4806]: I1127 10:31:42.247833 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" Nov 27 10:31:42 crc kubenswrapper[4806]: I1127 10:31:42.361805 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dd8e785-4b1d-406d-bb97-391b74415228-util\") pod \"4dd8e785-4b1d-406d-bb97-391b74415228\" (UID: \"4dd8e785-4b1d-406d-bb97-391b74415228\") " Nov 27 10:31:42 crc kubenswrapper[4806]: I1127 10:31:42.361887 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6v95f\" (UniqueName: \"kubernetes.io/projected/4dd8e785-4b1d-406d-bb97-391b74415228-kube-api-access-6v95f\") pod \"4dd8e785-4b1d-406d-bb97-391b74415228\" (UID: \"4dd8e785-4b1d-406d-bb97-391b74415228\") " Nov 27 10:31:42 crc kubenswrapper[4806]: I1127 10:31:42.361930 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dd8e785-4b1d-406d-bb97-391b74415228-bundle\") pod \"4dd8e785-4b1d-406d-bb97-391b74415228\" (UID: \"4dd8e785-4b1d-406d-bb97-391b74415228\") " Nov 27 10:31:42 crc kubenswrapper[4806]: I1127 10:31:42.363206 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dd8e785-4b1d-406d-bb97-391b74415228-bundle" (OuterVolumeSpecName: "bundle") pod "4dd8e785-4b1d-406d-bb97-391b74415228" (UID: "4dd8e785-4b1d-406d-bb97-391b74415228"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:31:42 crc kubenswrapper[4806]: I1127 10:31:42.373680 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dd8e785-4b1d-406d-bb97-391b74415228-kube-api-access-6v95f" (OuterVolumeSpecName: "kube-api-access-6v95f") pod "4dd8e785-4b1d-406d-bb97-391b74415228" (UID: "4dd8e785-4b1d-406d-bb97-391b74415228"). InnerVolumeSpecName "kube-api-access-6v95f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:31:42 crc kubenswrapper[4806]: I1127 10:31:42.382977 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dd8e785-4b1d-406d-bb97-391b74415228-util" (OuterVolumeSpecName: "util") pod "4dd8e785-4b1d-406d-bb97-391b74415228" (UID: "4dd8e785-4b1d-406d-bb97-391b74415228"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:31:42 crc kubenswrapper[4806]: I1127 10:31:42.463176 4806 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dd8e785-4b1d-406d-bb97-391b74415228-util\") on node \"crc\" DevicePath \"\"" Nov 27 10:31:42 crc kubenswrapper[4806]: I1127 10:31:42.463279 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6v95f\" (UniqueName: \"kubernetes.io/projected/4dd8e785-4b1d-406d-bb97-391b74415228-kube-api-access-6v95f\") on node \"crc\" DevicePath \"\"" Nov 27 10:31:42 crc kubenswrapper[4806]: I1127 10:31:42.463293 4806 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dd8e785-4b1d-406d-bb97-391b74415228-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:31:42 crc kubenswrapper[4806]: I1127 10:31:42.935967 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" event={"ID":"4dd8e785-4b1d-406d-bb97-391b74415228","Type":"ContainerDied","Data":"156b2534d1998f17fd945f399f687c4337480572a77e92ebc892e0846fad45b5"} Nov 27 10:31:42 crc kubenswrapper[4806]: I1127 10:31:42.936076 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="156b2534d1998f17fd945f399f687c4337480572a77e92ebc892e0846fad45b5" Nov 27 10:31:42 crc kubenswrapper[4806]: I1127 10:31:42.936108 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.206593 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-mtxjk"] Nov 27 10:31:44 crc kubenswrapper[4806]: E1127 10:31:44.207691 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dd8e785-4b1d-406d-bb97-391b74415228" containerName="pull" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.207755 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dd8e785-4b1d-406d-bb97-391b74415228" containerName="pull" Nov 27 10:31:44 crc kubenswrapper[4806]: E1127 10:31:44.207805 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dd8e785-4b1d-406d-bb97-391b74415228" containerName="util" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.207851 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dd8e785-4b1d-406d-bb97-391b74415228" containerName="util" Nov 27 10:31:44 crc kubenswrapper[4806]: E1127 10:31:44.207909 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dd8e785-4b1d-406d-bb97-391b74415228" containerName="extract" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.207956 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dd8e785-4b1d-406d-bb97-391b74415228" containerName="extract" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.208095 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dd8e785-4b1d-406d-bb97-391b74415228" containerName="extract" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.208536 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mtxjk" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.218828 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.224864 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-jtlbs" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.226104 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.232160 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-mtxjk"] Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.290334 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz587\" (UniqueName: \"kubernetes.io/projected/53a13ef4-b142-48f4-b6ae-81d8fb9a7203-kube-api-access-vz587\") pod \"nmstate-operator-5b5b58f5c8-mtxjk\" (UID: \"53a13ef4-b142-48f4-b6ae-81d8fb9a7203\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mtxjk" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.391856 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz587\" (UniqueName: \"kubernetes.io/projected/53a13ef4-b142-48f4-b6ae-81d8fb9a7203-kube-api-access-vz587\") pod \"nmstate-operator-5b5b58f5c8-mtxjk\" (UID: \"53a13ef4-b142-48f4-b6ae-81d8fb9a7203\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mtxjk" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.412528 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz587\" (UniqueName: \"kubernetes.io/projected/53a13ef4-b142-48f4-b6ae-81d8fb9a7203-kube-api-access-vz587\") pod \"nmstate-operator-5b5b58f5c8-mtxjk\" (UID: \"53a13ef4-b142-48f4-b6ae-81d8fb9a7203\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mtxjk" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.524525 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mtxjk" Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.746761 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-mtxjk"] Nov 27 10:31:44 crc kubenswrapper[4806]: I1127 10:31:44.947059 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mtxjk" event={"ID":"53a13ef4-b142-48f4-b6ae-81d8fb9a7203","Type":"ContainerStarted","Data":"4f08e17ba4ba665e72900f15fcc6849c79923f2a04a71a9f5f4c0b2f33d5f016"} Nov 27 10:31:48 crc kubenswrapper[4806]: I1127 10:31:48.996159 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mtxjk" event={"ID":"53a13ef4-b142-48f4-b6ae-81d8fb9a7203","Type":"ContainerStarted","Data":"b1799c58673f6657170b745658daf695c6622fba100f7430c4824c27042b6e8f"} Nov 27 10:31:49 crc kubenswrapper[4806]: I1127 10:31:49.017189 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mtxjk" podStartSLOduration=1.994510193 podStartE2EDuration="5.017165684s" podCreationTimestamp="2025-11-27 10:31:44 +0000 UTC" firstStartedPulling="2025-11-27 10:31:44.765083762 +0000 UTC m=+609.351674526" lastFinishedPulling="2025-11-27 10:31:47.787739253 +0000 UTC m=+612.374330017" observedRunningTime="2025-11-27 10:31:49.012118782 +0000 UTC m=+613.598709566" watchObservedRunningTime="2025-11-27 10:31:49.017165684 +0000 UTC m=+613.603756458" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.007975 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-np48l"] Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.008818 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-np48l" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.010980 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-jqzsz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.033642 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-np48l"] Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.037060 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq"] Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.037815 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.044994 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.065143 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq"] Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.069212 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lv9z\" (UniqueName: \"kubernetes.io/projected/412527c7-8920-4491-9035-2af48be2224b-kube-api-access-6lv9z\") pod \"nmstate-metrics-7f946cbc9-np48l\" (UID: \"412527c7-8920-4491-9035-2af48be2224b\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-np48l" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.069526 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/29deb9cb-7348-4dba-b3d3-cfbfdb88a79f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-lv2hq\" (UID: \"29deb9cb-7348-4dba-b3d3-cfbfdb88a79f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.069636 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc6bm\" (UniqueName: \"kubernetes.io/projected/29deb9cb-7348-4dba-b3d3-cfbfdb88a79f-kube-api-access-tc6bm\") pod \"nmstate-webhook-5f6d4c5ccb-lv2hq\" (UID: \"29deb9cb-7348-4dba-b3d3-cfbfdb88a79f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.103635 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-979gk"] Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.104365 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.170768 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lv9z\" (UniqueName: \"kubernetes.io/projected/412527c7-8920-4491-9035-2af48be2224b-kube-api-access-6lv9z\") pod \"nmstate-metrics-7f946cbc9-np48l\" (UID: \"412527c7-8920-4491-9035-2af48be2224b\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-np48l" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.170813 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/29deb9cb-7348-4dba-b3d3-cfbfdb88a79f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-lv2hq\" (UID: \"29deb9cb-7348-4dba-b3d3-cfbfdb88a79f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.170838 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/385b9d19-bcb8-40b8-ab88-7711ed6070e8-ovs-socket\") pod \"nmstate-handler-979gk\" (UID: \"385b9d19-bcb8-40b8-ab88-7711ed6070e8\") " pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.170855 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/385b9d19-bcb8-40b8-ab88-7711ed6070e8-dbus-socket\") pod \"nmstate-handler-979gk\" (UID: \"385b9d19-bcb8-40b8-ab88-7711ed6070e8\") " pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.170873 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/385b9d19-bcb8-40b8-ab88-7711ed6070e8-nmstate-lock\") pod \"nmstate-handler-979gk\" (UID: \"385b9d19-bcb8-40b8-ab88-7711ed6070e8\") " pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.170906 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc6bm\" (UniqueName: \"kubernetes.io/projected/29deb9cb-7348-4dba-b3d3-cfbfdb88a79f-kube-api-access-tc6bm\") pod \"nmstate-webhook-5f6d4c5ccb-lv2hq\" (UID: \"29deb9cb-7348-4dba-b3d3-cfbfdb88a79f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.170929 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzfql\" (UniqueName: \"kubernetes.io/projected/385b9d19-bcb8-40b8-ab88-7711ed6070e8-kube-api-access-bzfql\") pod \"nmstate-handler-979gk\" (UID: \"385b9d19-bcb8-40b8-ab88-7711ed6070e8\") " pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: E1127 10:31:50.171293 4806 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 27 10:31:50 crc kubenswrapper[4806]: E1127 10:31:50.171405 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29deb9cb-7348-4dba-b3d3-cfbfdb88a79f-tls-key-pair podName:29deb9cb-7348-4dba-b3d3-cfbfdb88a79f nodeName:}" failed. No retries permitted until 2025-11-27 10:31:50.671385926 +0000 UTC m=+615.257976690 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/29deb9cb-7348-4dba-b3d3-cfbfdb88a79f-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-lv2hq" (UID: "29deb9cb-7348-4dba-b3d3-cfbfdb88a79f") : secret "openshift-nmstate-webhook" not found Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.193198 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lv9z\" (UniqueName: \"kubernetes.io/projected/412527c7-8920-4491-9035-2af48be2224b-kube-api-access-6lv9z\") pod \"nmstate-metrics-7f946cbc9-np48l\" (UID: \"412527c7-8920-4491-9035-2af48be2224b\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-np48l" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.194289 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc6bm\" (UniqueName: \"kubernetes.io/projected/29deb9cb-7348-4dba-b3d3-cfbfdb88a79f-kube-api-access-tc6bm\") pod \"nmstate-webhook-5f6d4c5ccb-lv2hq\" (UID: \"29deb9cb-7348-4dba-b3d3-cfbfdb88a79f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.225252 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw"] Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.225964 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.230741 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.230748 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-qnw45" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.230786 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.250827 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw"] Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.271201 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzfql\" (UniqueName: \"kubernetes.io/projected/385b9d19-bcb8-40b8-ab88-7711ed6070e8-kube-api-access-bzfql\") pod \"nmstate-handler-979gk\" (UID: \"385b9d19-bcb8-40b8-ab88-7711ed6070e8\") " pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.271277 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/733b60f5-7297-420d-b952-896a3156f5f7-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-bncdw\" (UID: \"733b60f5-7297-420d-b952-896a3156f5f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.271316 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95tf2\" (UniqueName: \"kubernetes.io/projected/733b60f5-7297-420d-b952-896a3156f5f7-kube-api-access-95tf2\") pod \"nmstate-console-plugin-7fbb5f6569-bncdw\" (UID: \"733b60f5-7297-420d-b952-896a3156f5f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.271351 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/385b9d19-bcb8-40b8-ab88-7711ed6070e8-ovs-socket\") pod \"nmstate-handler-979gk\" (UID: \"385b9d19-bcb8-40b8-ab88-7711ed6070e8\") " pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.271368 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/385b9d19-bcb8-40b8-ab88-7711ed6070e8-dbus-socket\") pod \"nmstate-handler-979gk\" (UID: \"385b9d19-bcb8-40b8-ab88-7711ed6070e8\") " pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.271385 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/385b9d19-bcb8-40b8-ab88-7711ed6070e8-nmstate-lock\") pod \"nmstate-handler-979gk\" (UID: \"385b9d19-bcb8-40b8-ab88-7711ed6070e8\") " pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.271404 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/733b60f5-7297-420d-b952-896a3156f5f7-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-bncdw\" (UID: \"733b60f5-7297-420d-b952-896a3156f5f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.271739 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/385b9d19-bcb8-40b8-ab88-7711ed6070e8-ovs-socket\") pod \"nmstate-handler-979gk\" (UID: \"385b9d19-bcb8-40b8-ab88-7711ed6070e8\") " pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.272019 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/385b9d19-bcb8-40b8-ab88-7711ed6070e8-dbus-socket\") pod \"nmstate-handler-979gk\" (UID: \"385b9d19-bcb8-40b8-ab88-7711ed6070e8\") " pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.272053 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/385b9d19-bcb8-40b8-ab88-7711ed6070e8-nmstate-lock\") pod \"nmstate-handler-979gk\" (UID: \"385b9d19-bcb8-40b8-ab88-7711ed6070e8\") " pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.297659 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzfql\" (UniqueName: \"kubernetes.io/projected/385b9d19-bcb8-40b8-ab88-7711ed6070e8-kube-api-access-bzfql\") pod \"nmstate-handler-979gk\" (UID: \"385b9d19-bcb8-40b8-ab88-7711ed6070e8\") " pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.326364 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-np48l" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.372808 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/733b60f5-7297-420d-b952-896a3156f5f7-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-bncdw\" (UID: \"733b60f5-7297-420d-b952-896a3156f5f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.373203 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95tf2\" (UniqueName: \"kubernetes.io/projected/733b60f5-7297-420d-b952-896a3156f5f7-kube-api-access-95tf2\") pod \"nmstate-console-plugin-7fbb5f6569-bncdw\" (UID: \"733b60f5-7297-420d-b952-896a3156f5f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.373699 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/733b60f5-7297-420d-b952-896a3156f5f7-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-bncdw\" (UID: \"733b60f5-7297-420d-b952-896a3156f5f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" Nov 27 10:31:50 crc kubenswrapper[4806]: E1127 10:31:50.373843 4806 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.373973 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/733b60f5-7297-420d-b952-896a3156f5f7-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-bncdw\" (UID: \"733b60f5-7297-420d-b952-896a3156f5f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" Nov 27 10:31:50 crc kubenswrapper[4806]: E1127 10:31:50.374140 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/733b60f5-7297-420d-b952-896a3156f5f7-plugin-serving-cert podName:733b60f5-7297-420d-b952-896a3156f5f7 nodeName:}" failed. No retries permitted until 2025-11-27 10:31:50.874028265 +0000 UTC m=+615.460619039 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/733b60f5-7297-420d-b952-896a3156f5f7-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-bncdw" (UID: "733b60f5-7297-420d-b952-896a3156f5f7") : secret "plugin-serving-cert" not found Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.394768 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95tf2\" (UniqueName: \"kubernetes.io/projected/733b60f5-7297-420d-b952-896a3156f5f7-kube-api-access-95tf2\") pod \"nmstate-console-plugin-7fbb5f6569-bncdw\" (UID: \"733b60f5-7297-420d-b952-896a3156f5f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.424714 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.510373 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-77f96f7b98-sv5wz"] Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.511497 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.521031 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-77f96f7b98-sv5wz"] Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.576924 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fx6m\" (UniqueName: \"kubernetes.io/projected/a990bb25-5683-45d2-9165-3114f8f9d134-kube-api-access-6fx6m\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.577194 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a990bb25-5683-45d2-9165-3114f8f9d134-trusted-ca-bundle\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.577246 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a990bb25-5683-45d2-9165-3114f8f9d134-service-ca\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.577270 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a990bb25-5683-45d2-9165-3114f8f9d134-console-config\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.577290 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a990bb25-5683-45d2-9165-3114f8f9d134-console-oauth-config\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.577320 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a990bb25-5683-45d2-9165-3114f8f9d134-oauth-serving-cert\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.577356 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a990bb25-5683-45d2-9165-3114f8f9d134-console-serving-cert\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.673673 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-np48l"] Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.678624 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fx6m\" (UniqueName: \"kubernetes.io/projected/a990bb25-5683-45d2-9165-3114f8f9d134-kube-api-access-6fx6m\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.678685 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a990bb25-5683-45d2-9165-3114f8f9d134-trusted-ca-bundle\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.678724 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a990bb25-5683-45d2-9165-3114f8f9d134-service-ca\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.678745 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a990bb25-5683-45d2-9165-3114f8f9d134-console-config\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.678764 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a990bb25-5683-45d2-9165-3114f8f9d134-console-oauth-config\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.678796 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a990bb25-5683-45d2-9165-3114f8f9d134-oauth-serving-cert\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.678824 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/29deb9cb-7348-4dba-b3d3-cfbfdb88a79f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-lv2hq\" (UID: \"29deb9cb-7348-4dba-b3d3-cfbfdb88a79f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.678843 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a990bb25-5683-45d2-9165-3114f8f9d134-console-serving-cert\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.681576 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a990bb25-5683-45d2-9165-3114f8f9d134-trusted-ca-bundle\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.682196 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a990bb25-5683-45d2-9165-3114f8f9d134-service-ca\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.682820 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a990bb25-5683-45d2-9165-3114f8f9d134-console-config\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.687989 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a990bb25-5683-45d2-9165-3114f8f9d134-console-oauth-config\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.689483 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a990bb25-5683-45d2-9165-3114f8f9d134-oauth-serving-cert\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.689540 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a990bb25-5683-45d2-9165-3114f8f9d134-console-serving-cert\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.695035 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/29deb9cb-7348-4dba-b3d3-cfbfdb88a79f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-lv2hq\" (UID: \"29deb9cb-7348-4dba-b3d3-cfbfdb88a79f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.698130 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fx6m\" (UniqueName: \"kubernetes.io/projected/a990bb25-5683-45d2-9165-3114f8f9d134-kube-api-access-6fx6m\") pod \"console-77f96f7b98-sv5wz\" (UID: \"a990bb25-5683-45d2-9165-3114f8f9d134\") " pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.847756 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.884117 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/733b60f5-7297-420d-b952-896a3156f5f7-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-bncdw\" (UID: \"733b60f5-7297-420d-b952-896a3156f5f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.889325 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/733b60f5-7297-420d-b952-896a3156f5f7-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-bncdw\" (UID: \"733b60f5-7297-420d-b952-896a3156f5f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" Nov 27 10:31:50 crc kubenswrapper[4806]: I1127 10:31:50.951826 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" Nov 27 10:31:51 crc kubenswrapper[4806]: I1127 10:31:51.015922 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-979gk" event={"ID":"385b9d19-bcb8-40b8-ab88-7711ed6070e8","Type":"ContainerStarted","Data":"764e38d3a199ab5e4cbb295959f1a23ecf9a914acb75682bc8866d8d76b3f7ac"} Nov 27 10:31:51 crc kubenswrapper[4806]: I1127 10:31:51.023703 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-np48l" event={"ID":"412527c7-8920-4491-9035-2af48be2224b","Type":"ContainerStarted","Data":"d7265b5c1468595939a70f0fc9894b9755764bb5422b06d8099547856d1893b9"} Nov 27 10:31:51 crc kubenswrapper[4806]: I1127 10:31:51.057179 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-77f96f7b98-sv5wz"] Nov 27 10:31:51 crc kubenswrapper[4806]: I1127 10:31:51.155740 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" Nov 27 10:31:51 crc kubenswrapper[4806]: I1127 10:31:51.238197 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq"] Nov 27 10:31:51 crc kubenswrapper[4806]: W1127 10:31:51.250963 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29deb9cb_7348_4dba_b3d3_cfbfdb88a79f.slice/crio-f09fe4a5e0ed6eefc170fc7ae9c3b26f4f537e565c4a94e58d57332735159cd7 WatchSource:0}: Error finding container f09fe4a5e0ed6eefc170fc7ae9c3b26f4f537e565c4a94e58d57332735159cd7: Status 404 returned error can't find the container with id f09fe4a5e0ed6eefc170fc7ae9c3b26f4f537e565c4a94e58d57332735159cd7 Nov 27 10:31:51 crc kubenswrapper[4806]: I1127 10:31:51.448713 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw"] Nov 27 10:31:52 crc kubenswrapper[4806]: I1127 10:31:52.030709 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" event={"ID":"733b60f5-7297-420d-b952-896a3156f5f7","Type":"ContainerStarted","Data":"078b6adff60b40255ff63f8ee825b9a9fca417db61617a8dc4485b3694e392ae"} Nov 27 10:31:52 crc kubenswrapper[4806]: I1127 10:31:52.031479 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" event={"ID":"29deb9cb-7348-4dba-b3d3-cfbfdb88a79f","Type":"ContainerStarted","Data":"f09fe4a5e0ed6eefc170fc7ae9c3b26f4f537e565c4a94e58d57332735159cd7"} Nov 27 10:31:52 crc kubenswrapper[4806]: I1127 10:31:52.033812 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-77f96f7b98-sv5wz" event={"ID":"a990bb25-5683-45d2-9165-3114f8f9d134","Type":"ContainerStarted","Data":"9a24a8190ea5c8bedeba732929cb8459bfc32797e81ecde41d33f2b708aaa4c2"} Nov 27 10:31:52 crc kubenswrapper[4806]: I1127 10:31:52.033870 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-77f96f7b98-sv5wz" event={"ID":"a990bb25-5683-45d2-9165-3114f8f9d134","Type":"ContainerStarted","Data":"7fa184917270ea533c24863250b55705d338bcef3dbb21bca9c1980c440c8cbe"} Nov 27 10:31:52 crc kubenswrapper[4806]: I1127 10:31:52.057624 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-77f96f7b98-sv5wz" podStartSLOduration=2.057603778 podStartE2EDuration="2.057603778s" podCreationTimestamp="2025-11-27 10:31:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:31:52.053553063 +0000 UTC m=+616.640143827" watchObservedRunningTime="2025-11-27 10:31:52.057603778 +0000 UTC m=+616.644194542" Nov 27 10:31:54 crc kubenswrapper[4806]: I1127 10:31:54.054276 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" event={"ID":"29deb9cb-7348-4dba-b3d3-cfbfdb88a79f","Type":"ContainerStarted","Data":"2973e4ab11c142ca60f31a74b2ff947bd94971531f9cc993be7e22103f14fa4b"} Nov 27 10:31:54 crc kubenswrapper[4806]: I1127 10:31:54.055536 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" Nov 27 10:31:54 crc kubenswrapper[4806]: I1127 10:31:54.059137 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-979gk" event={"ID":"385b9d19-bcb8-40b8-ab88-7711ed6070e8","Type":"ContainerStarted","Data":"202be388016b8546dab17ac91f9d64758c97b7237fa74735cd11966be07b144d"} Nov 27 10:31:54 crc kubenswrapper[4806]: I1127 10:31:54.059371 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:31:54 crc kubenswrapper[4806]: I1127 10:31:54.061614 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-np48l" event={"ID":"412527c7-8920-4491-9035-2af48be2224b","Type":"ContainerStarted","Data":"a6f6bac82452a9b60530f4932af1e57f978f9022747ad22ea37d0aff1339bb12"} Nov 27 10:31:54 crc kubenswrapper[4806]: I1127 10:31:54.075637 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" podStartSLOduration=2.277585063 podStartE2EDuration="4.0756048s" podCreationTimestamp="2025-11-27 10:31:50 +0000 UTC" firstStartedPulling="2025-11-27 10:31:51.25961322 +0000 UTC m=+615.846203974" lastFinishedPulling="2025-11-27 10:31:53.057632947 +0000 UTC m=+617.644223711" observedRunningTime="2025-11-27 10:31:54.072538414 +0000 UTC m=+618.659129198" watchObservedRunningTime="2025-11-27 10:31:54.0756048 +0000 UTC m=+618.662195564" Nov 27 10:31:54 crc kubenswrapper[4806]: I1127 10:31:54.099935 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-979gk" podStartSLOduration=1.516558297 podStartE2EDuration="4.099905755s" podCreationTimestamp="2025-11-27 10:31:50 +0000 UTC" firstStartedPulling="2025-11-27 10:31:50.455451908 +0000 UTC m=+615.042042672" lastFinishedPulling="2025-11-27 10:31:53.038799366 +0000 UTC m=+617.625390130" observedRunningTime="2025-11-27 10:31:54.09972868 +0000 UTC m=+618.686319464" watchObservedRunningTime="2025-11-27 10:31:54.099905755 +0000 UTC m=+618.686496529" Nov 27 10:31:55 crc kubenswrapper[4806]: I1127 10:31:55.078359 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" event={"ID":"733b60f5-7297-420d-b952-896a3156f5f7","Type":"ContainerStarted","Data":"c2d111874f2e33a3615999abe4ce83320818be41c2746ea90b55f944e4468d51"} Nov 27 10:31:56 crc kubenswrapper[4806]: I1127 10:31:56.089691 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-np48l" event={"ID":"412527c7-8920-4491-9035-2af48be2224b","Type":"ContainerStarted","Data":"55da30b945ebc3b4f5e1a758e401e53a1a931dc300c6b414231761bb99962ffb"} Nov 27 10:31:56 crc kubenswrapper[4806]: I1127 10:31:56.112401 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-np48l" podStartSLOduration=2.233750481 podStartE2EDuration="7.112379942s" podCreationTimestamp="2025-11-27 10:31:49 +0000 UTC" firstStartedPulling="2025-11-27 10:31:50.693912045 +0000 UTC m=+615.280502809" lastFinishedPulling="2025-11-27 10:31:55.572541516 +0000 UTC m=+620.159132270" observedRunningTime="2025-11-27 10:31:56.105285772 +0000 UTC m=+620.691876536" watchObservedRunningTime="2025-11-27 10:31:56.112379942 +0000 UTC m=+620.698970706" Nov 27 10:31:56 crc kubenswrapper[4806]: I1127 10:31:56.125772 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bncdw" podStartSLOduration=3.134028758 podStartE2EDuration="6.125754689s" podCreationTimestamp="2025-11-27 10:31:50 +0000 UTC" firstStartedPulling="2025-11-27 10:31:51.47160067 +0000 UTC m=+616.058191434" lastFinishedPulling="2025-11-27 10:31:54.463326601 +0000 UTC m=+619.049917365" observedRunningTime="2025-11-27 10:31:55.099346567 +0000 UTC m=+619.685937331" watchObservedRunningTime="2025-11-27 10:31:56.125754689 +0000 UTC m=+620.712345453" Nov 27 10:32:00 crc kubenswrapper[4806]: I1127 10:32:00.451107 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-979gk" Nov 27 10:32:00 crc kubenswrapper[4806]: I1127 10:32:00.848163 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:32:00 crc kubenswrapper[4806]: I1127 10:32:00.848261 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:32:00 crc kubenswrapper[4806]: I1127 10:32:00.852927 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:32:01 crc kubenswrapper[4806]: I1127 10:32:01.123283 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-77f96f7b98-sv5wz" Nov 27 10:32:01 crc kubenswrapper[4806]: I1127 10:32:01.202936 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-6fqzk"] Nov 27 10:32:10 crc kubenswrapper[4806]: I1127 10:32:10.960133 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lv2hq" Nov 27 10:32:24 crc kubenswrapper[4806]: I1127 10:32:24.595830 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd"] Nov 27 10:32:24 crc kubenswrapper[4806]: I1127 10:32:24.599714 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" Nov 27 10:32:24 crc kubenswrapper[4806]: I1127 10:32:24.604157 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 27 10:32:24 crc kubenswrapper[4806]: I1127 10:32:24.882017 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd"] Nov 27 10:32:24 crc kubenswrapper[4806]: I1127 10:32:24.983583 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2afa20bc-c32f-4187-90b9-f41cf646cc8a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd\" (UID: \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" Nov 27 10:32:24 crc kubenswrapper[4806]: I1127 10:32:24.983774 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4gxl\" (UniqueName: \"kubernetes.io/projected/2afa20bc-c32f-4187-90b9-f41cf646cc8a-kube-api-access-s4gxl\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd\" (UID: \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" Nov 27 10:32:24 crc kubenswrapper[4806]: I1127 10:32:24.983926 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2afa20bc-c32f-4187-90b9-f41cf646cc8a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd\" (UID: \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" Nov 27 10:32:25 crc kubenswrapper[4806]: I1127 10:32:25.085360 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2afa20bc-c32f-4187-90b9-f41cf646cc8a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd\" (UID: \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" Nov 27 10:32:25 crc kubenswrapper[4806]: I1127 10:32:25.085858 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2afa20bc-c32f-4187-90b9-f41cf646cc8a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd\" (UID: \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" Nov 27 10:32:25 crc kubenswrapper[4806]: I1127 10:32:25.085959 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4gxl\" (UniqueName: \"kubernetes.io/projected/2afa20bc-c32f-4187-90b9-f41cf646cc8a-kube-api-access-s4gxl\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd\" (UID: \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" Nov 27 10:32:25 crc kubenswrapper[4806]: I1127 10:32:25.085957 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2afa20bc-c32f-4187-90b9-f41cf646cc8a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd\" (UID: \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" Nov 27 10:32:25 crc kubenswrapper[4806]: I1127 10:32:25.086284 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2afa20bc-c32f-4187-90b9-f41cf646cc8a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd\" (UID: \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" Nov 27 10:32:25 crc kubenswrapper[4806]: I1127 10:32:25.109731 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4gxl\" (UniqueName: \"kubernetes.io/projected/2afa20bc-c32f-4187-90b9-f41cf646cc8a-kube-api-access-s4gxl\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd\" (UID: \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" Nov 27 10:32:25 crc kubenswrapper[4806]: I1127 10:32:25.199384 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" Nov 27 10:32:25 crc kubenswrapper[4806]: I1127 10:32:25.661001 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd"] Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.248437 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-6fqzk" podUID="97042267-61ca-4864-a50a-60587abc7619" containerName="console" containerID="cri-o://10fb598b21604c336a6376d0e991967cd9fb1d34a7a4f7efe74b15d6921b0c8e" gracePeriod=15 Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.316005 4806 generic.go:334] "Generic (PLEG): container finished" podID="2afa20bc-c32f-4187-90b9-f41cf646cc8a" containerID="539211563a3a7c2a766a0f47bfd0f94f09916a79e17dba4cdef8ba3ee965d932" exitCode=0 Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.316323 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" event={"ID":"2afa20bc-c32f-4187-90b9-f41cf646cc8a","Type":"ContainerDied","Data":"539211563a3a7c2a766a0f47bfd0f94f09916a79e17dba4cdef8ba3ee965d932"} Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.316484 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" event={"ID":"2afa20bc-c32f-4187-90b9-f41cf646cc8a","Type":"ContainerStarted","Data":"fbc93b9e9d89bcd17a26147ca125fbc2915b5ccc96a6c9fc059e7c3e710b7af4"} Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.739953 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-6fqzk_97042267-61ca-4864-a50a-60587abc7619/console/0.log" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.740037 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.813501 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-console-config\") pod \"97042267-61ca-4864-a50a-60587abc7619\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.813558 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/97042267-61ca-4864-a50a-60587abc7619-console-serving-cert\") pod \"97042267-61ca-4864-a50a-60587abc7619\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.813585 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-service-ca\") pod \"97042267-61ca-4864-a50a-60587abc7619\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.813645 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/97042267-61ca-4864-a50a-60587abc7619-console-oauth-config\") pod \"97042267-61ca-4864-a50a-60587abc7619\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.813673 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7dtd\" (UniqueName: \"kubernetes.io/projected/97042267-61ca-4864-a50a-60587abc7619-kube-api-access-j7dtd\") pod \"97042267-61ca-4864-a50a-60587abc7619\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.813695 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-oauth-serving-cert\") pod \"97042267-61ca-4864-a50a-60587abc7619\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.813712 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-trusted-ca-bundle\") pod \"97042267-61ca-4864-a50a-60587abc7619\" (UID: \"97042267-61ca-4864-a50a-60587abc7619\") " Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.815138 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-service-ca" (OuterVolumeSpecName: "service-ca") pod "97042267-61ca-4864-a50a-60587abc7619" (UID: "97042267-61ca-4864-a50a-60587abc7619"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.815356 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-console-config" (OuterVolumeSpecName: "console-config") pod "97042267-61ca-4864-a50a-60587abc7619" (UID: "97042267-61ca-4864-a50a-60587abc7619"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.815410 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "97042267-61ca-4864-a50a-60587abc7619" (UID: "97042267-61ca-4864-a50a-60587abc7619"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.815629 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "97042267-61ca-4864-a50a-60587abc7619" (UID: "97042267-61ca-4864-a50a-60587abc7619"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.820564 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97042267-61ca-4864-a50a-60587abc7619-kube-api-access-j7dtd" (OuterVolumeSpecName: "kube-api-access-j7dtd") pod "97042267-61ca-4864-a50a-60587abc7619" (UID: "97042267-61ca-4864-a50a-60587abc7619"). InnerVolumeSpecName "kube-api-access-j7dtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.822657 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97042267-61ca-4864-a50a-60587abc7619-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "97042267-61ca-4864-a50a-60587abc7619" (UID: "97042267-61ca-4864-a50a-60587abc7619"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.823739 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97042267-61ca-4864-a50a-60587abc7619-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "97042267-61ca-4864-a50a-60587abc7619" (UID: "97042267-61ca-4864-a50a-60587abc7619"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.914800 4806 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/97042267-61ca-4864-a50a-60587abc7619-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.915107 4806 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-console-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.915165 4806 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-service-ca\") on node \"crc\" DevicePath \"\"" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.915215 4806 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/97042267-61ca-4864-a50a-60587abc7619-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.915412 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7dtd\" (UniqueName: \"kubernetes.io/projected/97042267-61ca-4864-a50a-60587abc7619-kube-api-access-j7dtd\") on node \"crc\" DevicePath \"\"" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.915483 4806 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 27 10:32:26 crc kubenswrapper[4806]: I1127 10:32:26.915547 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97042267-61ca-4864-a50a-60587abc7619-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:32:27 crc kubenswrapper[4806]: I1127 10:32:27.325011 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-6fqzk_97042267-61ca-4864-a50a-60587abc7619/console/0.log" Nov 27 10:32:27 crc kubenswrapper[4806]: I1127 10:32:27.325355 4806 generic.go:334] "Generic (PLEG): container finished" podID="97042267-61ca-4864-a50a-60587abc7619" containerID="10fb598b21604c336a6376d0e991967cd9fb1d34a7a4f7efe74b15d6921b0c8e" exitCode=2 Nov 27 10:32:27 crc kubenswrapper[4806]: I1127 10:32:27.325389 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6fqzk" event={"ID":"97042267-61ca-4864-a50a-60587abc7619","Type":"ContainerDied","Data":"10fb598b21604c336a6376d0e991967cd9fb1d34a7a4f7efe74b15d6921b0c8e"} Nov 27 10:32:27 crc kubenswrapper[4806]: I1127 10:32:27.325411 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6fqzk" Nov 27 10:32:27 crc kubenswrapper[4806]: I1127 10:32:27.325423 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6fqzk" event={"ID":"97042267-61ca-4864-a50a-60587abc7619","Type":"ContainerDied","Data":"46651de76bba572654753c832084eb28b0db7ecba5777cad0076c1426de15951"} Nov 27 10:32:27 crc kubenswrapper[4806]: I1127 10:32:27.325460 4806 scope.go:117] "RemoveContainer" containerID="10fb598b21604c336a6376d0e991967cd9fb1d34a7a4f7efe74b15d6921b0c8e" Nov 27 10:32:27 crc kubenswrapper[4806]: I1127 10:32:27.359528 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-6fqzk"] Nov 27 10:32:27 crc kubenswrapper[4806]: I1127 10:32:27.366321 4806 scope.go:117] "RemoveContainer" containerID="10fb598b21604c336a6376d0e991967cd9fb1d34a7a4f7efe74b15d6921b0c8e" Nov 27 10:32:27 crc kubenswrapper[4806]: E1127 10:32:27.367187 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10fb598b21604c336a6376d0e991967cd9fb1d34a7a4f7efe74b15d6921b0c8e\": container with ID starting with 10fb598b21604c336a6376d0e991967cd9fb1d34a7a4f7efe74b15d6921b0c8e not found: ID does not exist" containerID="10fb598b21604c336a6376d0e991967cd9fb1d34a7a4f7efe74b15d6921b0c8e" Nov 27 10:32:27 crc kubenswrapper[4806]: I1127 10:32:27.367240 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10fb598b21604c336a6376d0e991967cd9fb1d34a7a4f7efe74b15d6921b0c8e"} err="failed to get container status \"10fb598b21604c336a6376d0e991967cd9fb1d34a7a4f7efe74b15d6921b0c8e\": rpc error: code = NotFound desc = could not find container \"10fb598b21604c336a6376d0e991967cd9fb1d34a7a4f7efe74b15d6921b0c8e\": container with ID starting with 10fb598b21604c336a6376d0e991967cd9fb1d34a7a4f7efe74b15d6921b0c8e not found: ID does not exist" Nov 27 10:32:27 crc kubenswrapper[4806]: I1127 10:32:27.371174 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-6fqzk"] Nov 27 10:32:28 crc kubenswrapper[4806]: I1127 10:32:28.125522 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97042267-61ca-4864-a50a-60587abc7619" path="/var/lib/kubelet/pods/97042267-61ca-4864-a50a-60587abc7619/volumes" Nov 27 10:32:28 crc kubenswrapper[4806]: I1127 10:32:28.336701 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" event={"ID":"2afa20bc-c32f-4187-90b9-f41cf646cc8a","Type":"ContainerDied","Data":"ed19d913d2a79a6025c48517e23b2bd3d6d2d33a4c66e6b9fc1ef6957183a3bf"} Nov 27 10:32:28 crc kubenswrapper[4806]: I1127 10:32:28.336802 4806 generic.go:334] "Generic (PLEG): container finished" podID="2afa20bc-c32f-4187-90b9-f41cf646cc8a" containerID="ed19d913d2a79a6025c48517e23b2bd3d6d2d33a4c66e6b9fc1ef6957183a3bf" exitCode=0 Nov 27 10:32:29 crc kubenswrapper[4806]: I1127 10:32:29.350647 4806 generic.go:334] "Generic (PLEG): container finished" podID="2afa20bc-c32f-4187-90b9-f41cf646cc8a" containerID="7a7babfd56e4c0e50ab7e44597f8bf61b19e71ef4487d891dbf1e7cb0156ec1c" exitCode=0 Nov 27 10:32:29 crc kubenswrapper[4806]: I1127 10:32:29.350714 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" event={"ID":"2afa20bc-c32f-4187-90b9-f41cf646cc8a","Type":"ContainerDied","Data":"7a7babfd56e4c0e50ab7e44597f8bf61b19e71ef4487d891dbf1e7cb0156ec1c"} Nov 27 10:32:30 crc kubenswrapper[4806]: I1127 10:32:30.596596 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" Nov 27 10:32:30 crc kubenswrapper[4806]: I1127 10:32:30.671040 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4gxl\" (UniqueName: \"kubernetes.io/projected/2afa20bc-c32f-4187-90b9-f41cf646cc8a-kube-api-access-s4gxl\") pod \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\" (UID: \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\") " Nov 27 10:32:30 crc kubenswrapper[4806]: I1127 10:32:30.671114 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2afa20bc-c32f-4187-90b9-f41cf646cc8a-bundle\") pod \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\" (UID: \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\") " Nov 27 10:32:30 crc kubenswrapper[4806]: I1127 10:32:30.671161 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2afa20bc-c32f-4187-90b9-f41cf646cc8a-util\") pod \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\" (UID: \"2afa20bc-c32f-4187-90b9-f41cf646cc8a\") " Nov 27 10:32:30 crc kubenswrapper[4806]: I1127 10:32:30.672182 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2afa20bc-c32f-4187-90b9-f41cf646cc8a-bundle" (OuterVolumeSpecName: "bundle") pod "2afa20bc-c32f-4187-90b9-f41cf646cc8a" (UID: "2afa20bc-c32f-4187-90b9-f41cf646cc8a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:32:30 crc kubenswrapper[4806]: I1127 10:32:30.677594 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2afa20bc-c32f-4187-90b9-f41cf646cc8a-kube-api-access-s4gxl" (OuterVolumeSpecName: "kube-api-access-s4gxl") pod "2afa20bc-c32f-4187-90b9-f41cf646cc8a" (UID: "2afa20bc-c32f-4187-90b9-f41cf646cc8a"). InnerVolumeSpecName "kube-api-access-s4gxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:32:30 crc kubenswrapper[4806]: I1127 10:32:30.685136 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2afa20bc-c32f-4187-90b9-f41cf646cc8a-util" (OuterVolumeSpecName: "util") pod "2afa20bc-c32f-4187-90b9-f41cf646cc8a" (UID: "2afa20bc-c32f-4187-90b9-f41cf646cc8a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:32:30 crc kubenswrapper[4806]: I1127 10:32:30.772493 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4gxl\" (UniqueName: \"kubernetes.io/projected/2afa20bc-c32f-4187-90b9-f41cf646cc8a-kube-api-access-s4gxl\") on node \"crc\" DevicePath \"\"" Nov 27 10:32:30 crc kubenswrapper[4806]: I1127 10:32:30.772531 4806 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2afa20bc-c32f-4187-90b9-f41cf646cc8a-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:32:30 crc kubenswrapper[4806]: I1127 10:32:30.772543 4806 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2afa20bc-c32f-4187-90b9-f41cf646cc8a-util\") on node \"crc\" DevicePath \"\"" Nov 27 10:32:31 crc kubenswrapper[4806]: I1127 10:32:31.366720 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" event={"ID":"2afa20bc-c32f-4187-90b9-f41cf646cc8a","Type":"ContainerDied","Data":"fbc93b9e9d89bcd17a26147ca125fbc2915b5ccc96a6c9fc059e7c3e710b7af4"} Nov 27 10:32:31 crc kubenswrapper[4806]: I1127 10:32:31.366762 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd" Nov 27 10:32:31 crc kubenswrapper[4806]: I1127 10:32:31.366775 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbc93b9e9d89bcd17a26147ca125fbc2915b5ccc96a6c9fc059e7c3e710b7af4" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.644231 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk"] Nov 27 10:32:42 crc kubenswrapper[4806]: E1127 10:32:42.645177 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2afa20bc-c32f-4187-90b9-f41cf646cc8a" containerName="util" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.645193 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2afa20bc-c32f-4187-90b9-f41cf646cc8a" containerName="util" Nov 27 10:32:42 crc kubenswrapper[4806]: E1127 10:32:42.645203 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2afa20bc-c32f-4187-90b9-f41cf646cc8a" containerName="extract" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.645227 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2afa20bc-c32f-4187-90b9-f41cf646cc8a" containerName="extract" Nov 27 10:32:42 crc kubenswrapper[4806]: E1127 10:32:42.645249 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97042267-61ca-4864-a50a-60587abc7619" containerName="console" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.645256 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="97042267-61ca-4864-a50a-60587abc7619" containerName="console" Nov 27 10:32:42 crc kubenswrapper[4806]: E1127 10:32:42.645273 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2afa20bc-c32f-4187-90b9-f41cf646cc8a" containerName="pull" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.645280 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2afa20bc-c32f-4187-90b9-f41cf646cc8a" containerName="pull" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.645494 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="2afa20bc-c32f-4187-90b9-f41cf646cc8a" containerName="extract" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.645533 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="97042267-61ca-4864-a50a-60587abc7619" containerName="console" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.646201 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.649765 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.650163 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.650405 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.650574 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.650700 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-mmkp2" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.666913 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk"] Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.750492 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxzf2\" (UniqueName: \"kubernetes.io/projected/d62398ac-46b4-4e4f-b3f9-9ae58756b4b3-kube-api-access-kxzf2\") pod \"metallb-operator-controller-manager-766c64499f-pl7sk\" (UID: \"d62398ac-46b4-4e4f-b3f9-9ae58756b4b3\") " pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.750567 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d62398ac-46b4-4e4f-b3f9-9ae58756b4b3-apiservice-cert\") pod \"metallb-operator-controller-manager-766c64499f-pl7sk\" (UID: \"d62398ac-46b4-4e4f-b3f9-9ae58756b4b3\") " pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.750642 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d62398ac-46b4-4e4f-b3f9-9ae58756b4b3-webhook-cert\") pod \"metallb-operator-controller-manager-766c64499f-pl7sk\" (UID: \"d62398ac-46b4-4e4f-b3f9-9ae58756b4b3\") " pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.851670 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxzf2\" (UniqueName: \"kubernetes.io/projected/d62398ac-46b4-4e4f-b3f9-9ae58756b4b3-kube-api-access-kxzf2\") pod \"metallb-operator-controller-manager-766c64499f-pl7sk\" (UID: \"d62398ac-46b4-4e4f-b3f9-9ae58756b4b3\") " pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.851728 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d62398ac-46b4-4e4f-b3f9-9ae58756b4b3-apiservice-cert\") pod \"metallb-operator-controller-manager-766c64499f-pl7sk\" (UID: \"d62398ac-46b4-4e4f-b3f9-9ae58756b4b3\") " pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.851781 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d62398ac-46b4-4e4f-b3f9-9ae58756b4b3-webhook-cert\") pod \"metallb-operator-controller-manager-766c64499f-pl7sk\" (UID: \"d62398ac-46b4-4e4f-b3f9-9ae58756b4b3\") " pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.861079 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d62398ac-46b4-4e4f-b3f9-9ae58756b4b3-webhook-cert\") pod \"metallb-operator-controller-manager-766c64499f-pl7sk\" (UID: \"d62398ac-46b4-4e4f-b3f9-9ae58756b4b3\") " pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.861091 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d62398ac-46b4-4e4f-b3f9-9ae58756b4b3-apiservice-cert\") pod \"metallb-operator-controller-manager-766c64499f-pl7sk\" (UID: \"d62398ac-46b4-4e4f-b3f9-9ae58756b4b3\") " pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.888583 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxzf2\" (UniqueName: \"kubernetes.io/projected/d62398ac-46b4-4e4f-b3f9-9ae58756b4b3-kube-api-access-kxzf2\") pod \"metallb-operator-controller-manager-766c64499f-pl7sk\" (UID: \"d62398ac-46b4-4e4f-b3f9-9ae58756b4b3\") " pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.962369 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.987722 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf"] Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.989155 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.993211 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.993464 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 27 10:32:42 crc kubenswrapper[4806]: I1127 10:32:42.993539 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-cmn7r" Nov 27 10:32:43 crc kubenswrapper[4806]: I1127 10:32:43.022114 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf"] Nov 27 10:32:43 crc kubenswrapper[4806]: I1127 10:32:43.155335 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/020fead9-ad47-4930-8797-88254cbc0245-webhook-cert\") pod \"metallb-operator-webhook-server-7cbfdf6f8b-zwpmf\" (UID: \"020fead9-ad47-4930-8797-88254cbc0245\") " pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" Nov 27 10:32:43 crc kubenswrapper[4806]: I1127 10:32:43.155405 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdsfd\" (UniqueName: \"kubernetes.io/projected/020fead9-ad47-4930-8797-88254cbc0245-kube-api-access-wdsfd\") pod \"metallb-operator-webhook-server-7cbfdf6f8b-zwpmf\" (UID: \"020fead9-ad47-4930-8797-88254cbc0245\") " pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" Nov 27 10:32:43 crc kubenswrapper[4806]: I1127 10:32:43.155445 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/020fead9-ad47-4930-8797-88254cbc0245-apiservice-cert\") pod \"metallb-operator-webhook-server-7cbfdf6f8b-zwpmf\" (UID: \"020fead9-ad47-4930-8797-88254cbc0245\") " pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" Nov 27 10:32:43 crc kubenswrapper[4806]: I1127 10:32:43.257929 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdsfd\" (UniqueName: \"kubernetes.io/projected/020fead9-ad47-4930-8797-88254cbc0245-kube-api-access-wdsfd\") pod \"metallb-operator-webhook-server-7cbfdf6f8b-zwpmf\" (UID: \"020fead9-ad47-4930-8797-88254cbc0245\") " pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" Nov 27 10:32:43 crc kubenswrapper[4806]: I1127 10:32:43.258472 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/020fead9-ad47-4930-8797-88254cbc0245-apiservice-cert\") pod \"metallb-operator-webhook-server-7cbfdf6f8b-zwpmf\" (UID: \"020fead9-ad47-4930-8797-88254cbc0245\") " pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" Nov 27 10:32:43 crc kubenswrapper[4806]: I1127 10:32:43.258544 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/020fead9-ad47-4930-8797-88254cbc0245-webhook-cert\") pod \"metallb-operator-webhook-server-7cbfdf6f8b-zwpmf\" (UID: \"020fead9-ad47-4930-8797-88254cbc0245\") " pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" Nov 27 10:32:43 crc kubenswrapper[4806]: I1127 10:32:43.263448 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/020fead9-ad47-4930-8797-88254cbc0245-webhook-cert\") pod \"metallb-operator-webhook-server-7cbfdf6f8b-zwpmf\" (UID: \"020fead9-ad47-4930-8797-88254cbc0245\") " pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" Nov 27 10:32:43 crc kubenswrapper[4806]: I1127 10:32:43.273621 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/020fead9-ad47-4930-8797-88254cbc0245-apiservice-cert\") pod \"metallb-operator-webhook-server-7cbfdf6f8b-zwpmf\" (UID: \"020fead9-ad47-4930-8797-88254cbc0245\") " pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" Nov 27 10:32:43 crc kubenswrapper[4806]: I1127 10:32:43.303027 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdsfd\" (UniqueName: \"kubernetes.io/projected/020fead9-ad47-4930-8797-88254cbc0245-kube-api-access-wdsfd\") pod \"metallb-operator-webhook-server-7cbfdf6f8b-zwpmf\" (UID: \"020fead9-ad47-4930-8797-88254cbc0245\") " pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" Nov 27 10:32:43 crc kubenswrapper[4806]: I1127 10:32:43.343512 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" Nov 27 10:32:43 crc kubenswrapper[4806]: I1127 10:32:43.522443 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk"] Nov 27 10:32:43 crc kubenswrapper[4806]: W1127 10:32:43.538388 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd62398ac_46b4_4e4f_b3f9_9ae58756b4b3.slice/crio-3e1ed0d3f18036be89f2ad1cfe95379dba4d975d39456d0ad8371b2372916974 WatchSource:0}: Error finding container 3e1ed0d3f18036be89f2ad1cfe95379dba4d975d39456d0ad8371b2372916974: Status 404 returned error can't find the container with id 3e1ed0d3f18036be89f2ad1cfe95379dba4d975d39456d0ad8371b2372916974 Nov 27 10:32:43 crc kubenswrapper[4806]: I1127 10:32:43.782522 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf"] Nov 27 10:32:43 crc kubenswrapper[4806]: W1127 10:32:43.793062 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod020fead9_ad47_4930_8797_88254cbc0245.slice/crio-8024fd5b6bdefffbf5710d5d59db186844ba41a889209c556a3db46620bead82 WatchSource:0}: Error finding container 8024fd5b6bdefffbf5710d5d59db186844ba41a889209c556a3db46620bead82: Status 404 returned error can't find the container with id 8024fd5b6bdefffbf5710d5d59db186844ba41a889209c556a3db46620bead82 Nov 27 10:32:44 crc kubenswrapper[4806]: I1127 10:32:44.463387 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" event={"ID":"d62398ac-46b4-4e4f-b3f9-9ae58756b4b3","Type":"ContainerStarted","Data":"3e1ed0d3f18036be89f2ad1cfe95379dba4d975d39456d0ad8371b2372916974"} Nov 27 10:32:44 crc kubenswrapper[4806]: I1127 10:32:44.464819 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" event={"ID":"020fead9-ad47-4930-8797-88254cbc0245","Type":"ContainerStarted","Data":"8024fd5b6bdefffbf5710d5d59db186844ba41a889209c556a3db46620bead82"} Nov 27 10:32:50 crc kubenswrapper[4806]: I1127 10:32:50.512452 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" event={"ID":"020fead9-ad47-4930-8797-88254cbc0245","Type":"ContainerStarted","Data":"81d3ec6b488165ab161ea15f3ba8f5fba7e52a0609db9e9c9ad2c5511fa10cdf"} Nov 27 10:32:50 crc kubenswrapper[4806]: I1127 10:32:50.513752 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" Nov 27 10:32:50 crc kubenswrapper[4806]: I1127 10:32:50.515895 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" event={"ID":"d62398ac-46b4-4e4f-b3f9-9ae58756b4b3","Type":"ContainerStarted","Data":"034bca509c7a4203bebc5dcee6c9b732ee931a0c2adba3919e65369c20f14045"} Nov 27 10:32:50 crc kubenswrapper[4806]: I1127 10:32:50.516285 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" Nov 27 10:32:50 crc kubenswrapper[4806]: I1127 10:32:50.535203 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" podStartSLOduration=2.792902109 podStartE2EDuration="8.535187157s" podCreationTimestamp="2025-11-27 10:32:42 +0000 UTC" firstStartedPulling="2025-11-27 10:32:43.796722469 +0000 UTC m=+668.383313233" lastFinishedPulling="2025-11-27 10:32:49.539007527 +0000 UTC m=+674.125598281" observedRunningTime="2025-11-27 10:32:50.531106763 +0000 UTC m=+675.117697547" watchObservedRunningTime="2025-11-27 10:32:50.535187157 +0000 UTC m=+675.121777931" Nov 27 10:32:50 crc kubenswrapper[4806]: I1127 10:32:50.551380 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" podStartSLOduration=2.583171691 podStartE2EDuration="8.551364953s" podCreationTimestamp="2025-11-27 10:32:42 +0000 UTC" firstStartedPulling="2025-11-27 10:32:43.549411042 +0000 UTC m=+668.136001806" lastFinishedPulling="2025-11-27 10:32:49.517604304 +0000 UTC m=+674.104195068" observedRunningTime="2025-11-27 10:32:50.55052893 +0000 UTC m=+675.137119704" watchObservedRunningTime="2025-11-27 10:32:50.551364953 +0000 UTC m=+675.137955727" Nov 27 10:33:03 crc kubenswrapper[4806]: I1127 10:33:03.349637 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7cbfdf6f8b-zwpmf" Nov 27 10:33:14 crc kubenswrapper[4806]: I1127 10:33:14.451097 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:33:14 crc kubenswrapper[4806]: I1127 10:33:14.451769 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:33:22 crc kubenswrapper[4806]: I1127 10:33:22.966631 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-766c64499f-pl7sk" Nov 27 10:33:23 crc kubenswrapper[4806]: I1127 10:33:23.899279 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-hd27h"] Nov 27 10:33:23 crc kubenswrapper[4806]: I1127 10:33:23.901445 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:23 crc kubenswrapper[4806]: I1127 10:33:23.906780 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2"] Nov 27 10:33:23 crc kubenswrapper[4806]: I1127 10:33:23.907800 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 27 10:33:23 crc kubenswrapper[4806]: I1127 10:33:23.907895 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 27 10:33:23 crc kubenswrapper[4806]: I1127 10:33:23.908640 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" Nov 27 10:33:23 crc kubenswrapper[4806]: I1127 10:33:23.911560 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 27 10:33:23 crc kubenswrapper[4806]: I1127 10:33:23.924391 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2"] Nov 27 10:33:23 crc kubenswrapper[4806]: I1127 10:33:23.927895 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-nf9pb" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.010131 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhpj8\" (UniqueName: \"kubernetes.io/projected/5e7e1847-fefb-4289-8dbb-c04dec670021-kube-api-access-vhpj8\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.010192 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5e7e1847-fefb-4289-8dbb-c04dec670021-reloader\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.010251 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5e7e1847-fefb-4289-8dbb-c04dec670021-metrics\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.010276 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bkgj\" (UniqueName: \"kubernetes.io/projected/ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b-kube-api-access-8bkgj\") pod \"frr-k8s-webhook-server-7fcb986d4-l4ld2\" (UID: \"ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.010298 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7e1847-fefb-4289-8dbb-c04dec670021-metrics-certs\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.010317 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-l4ld2\" (UID: \"ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.010344 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5e7e1847-fefb-4289-8dbb-c04dec670021-frr-startup\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.010368 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5e7e1847-fefb-4289-8dbb-c04dec670021-frr-conf\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.010392 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5e7e1847-fefb-4289-8dbb-c04dec670021-frr-sockets\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.042205 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-bvd28"] Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.043372 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bvd28" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.047746 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.047802 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.048109 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.048754 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-tknbp" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.112418 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bkgj\" (UniqueName: \"kubernetes.io/projected/ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b-kube-api-access-8bkgj\") pod \"frr-k8s-webhook-server-7fcb986d4-l4ld2\" (UID: \"ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.112463 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7e1847-fefb-4289-8dbb-c04dec670021-metrics-certs\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.112485 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-l4ld2\" (UID: \"ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.112519 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5e7e1847-fefb-4289-8dbb-c04dec670021-frr-startup\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.112540 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5e7e1847-fefb-4289-8dbb-c04dec670021-frr-conf\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.112565 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5e7e1847-fefb-4289-8dbb-c04dec670021-frr-sockets\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.112619 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhpj8\" (UniqueName: \"kubernetes.io/projected/5e7e1847-fefb-4289-8dbb-c04dec670021-kube-api-access-vhpj8\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: E1127 10:33:24.113176 4806 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 27 10:33:24 crc kubenswrapper[4806]: E1127 10:33:24.113252 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e7e1847-fefb-4289-8dbb-c04dec670021-metrics-certs podName:5e7e1847-fefb-4289-8dbb-c04dec670021 nodeName:}" failed. No retries permitted until 2025-11-27 10:33:24.613216661 +0000 UTC m=+709.199807425 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5e7e1847-fefb-4289-8dbb-c04dec670021-metrics-certs") pod "frr-k8s-hd27h" (UID: "5e7e1847-fefb-4289-8dbb-c04dec670021") : secret "frr-k8s-certs-secret" not found Nov 27 10:33:24 crc kubenswrapper[4806]: E1127 10:33:24.113429 4806 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 27 10:33:24 crc kubenswrapper[4806]: E1127 10:33:24.113459 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b-cert podName:ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b nodeName:}" failed. No retries permitted until 2025-11-27 10:33:24.613451888 +0000 UTC m=+709.200042652 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b-cert") pod "frr-k8s-webhook-server-7fcb986d4-l4ld2" (UID: "ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b") : secret "frr-k8s-webhook-server-cert" not found Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.114552 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5e7e1847-fefb-4289-8dbb-c04dec670021-frr-sockets\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.114628 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5e7e1847-fefb-4289-8dbb-c04dec670021-reloader\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.114653 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5e7e1847-fefb-4289-8dbb-c04dec670021-metrics\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.114861 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5e7e1847-fefb-4289-8dbb-c04dec670021-frr-conf\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.114953 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5e7e1847-fefb-4289-8dbb-c04dec670021-reloader\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.115136 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5e7e1847-fefb-4289-8dbb-c04dec670021-metrics\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.115263 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5e7e1847-fefb-4289-8dbb-c04dec670021-frr-startup\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.143938 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-7c7qd"] Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.145010 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-7c7qd" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.148565 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.199396 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-7c7qd"] Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.201616 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhpj8\" (UniqueName: \"kubernetes.io/projected/5e7e1847-fefb-4289-8dbb-c04dec670021-kube-api-access-vhpj8\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.220762 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-metrics-certs\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.220848 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgtsq\" (UniqueName: \"kubernetes.io/projected/ec562de0-63d3-4613-96f6-b44435bd7dab-kube-api-access-hgtsq\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.220899 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ec562de0-63d3-4613-96f6-b44435bd7dab-metallb-excludel2\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.221051 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-memberlist\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.233201 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bkgj\" (UniqueName: \"kubernetes.io/projected/ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b-kube-api-access-8bkgj\") pod \"frr-k8s-webhook-server-7fcb986d4-l4ld2\" (UID: \"ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.321947 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d73ced61-9b8c-475c-b971-362279798b8e-metrics-certs\") pod \"controller-f8648f98b-7c7qd\" (UID: \"d73ced61-9b8c-475c-b971-362279798b8e\") " pod="metallb-system/controller-f8648f98b-7c7qd" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.322048 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d73ced61-9b8c-475c-b971-362279798b8e-cert\") pod \"controller-f8648f98b-7c7qd\" (UID: \"d73ced61-9b8c-475c-b971-362279798b8e\") " pod="metallb-system/controller-f8648f98b-7c7qd" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.322071 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-memberlist\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.322097 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-metrics-certs\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.322125 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgtsq\" (UniqueName: \"kubernetes.io/projected/ec562de0-63d3-4613-96f6-b44435bd7dab-kube-api-access-hgtsq\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.322143 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t7vr\" (UniqueName: \"kubernetes.io/projected/d73ced61-9b8c-475c-b971-362279798b8e-kube-api-access-8t7vr\") pod \"controller-f8648f98b-7c7qd\" (UID: \"d73ced61-9b8c-475c-b971-362279798b8e\") " pod="metallb-system/controller-f8648f98b-7c7qd" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.322165 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ec562de0-63d3-4613-96f6-b44435bd7dab-metallb-excludel2\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:24 crc kubenswrapper[4806]: E1127 10:33:24.322404 4806 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 27 10:33:24 crc kubenswrapper[4806]: E1127 10:33:24.322497 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-memberlist podName:ec562de0-63d3-4613-96f6-b44435bd7dab nodeName:}" failed. No retries permitted until 2025-11-27 10:33:24.822470246 +0000 UTC m=+709.409061010 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-memberlist") pod "speaker-bvd28" (UID: "ec562de0-63d3-4613-96f6-b44435bd7dab") : secret "metallb-memberlist" not found Nov 27 10:33:24 crc kubenswrapper[4806]: E1127 10:33:24.322801 4806 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Nov 27 10:33:24 crc kubenswrapper[4806]: E1127 10:33:24.322834 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-metrics-certs podName:ec562de0-63d3-4613-96f6-b44435bd7dab nodeName:}" failed. No retries permitted until 2025-11-27 10:33:24.822827306 +0000 UTC m=+709.409418070 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-metrics-certs") pod "speaker-bvd28" (UID: "ec562de0-63d3-4613-96f6-b44435bd7dab") : secret "speaker-certs-secret" not found Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.322921 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ec562de0-63d3-4613-96f6-b44435bd7dab-metallb-excludel2\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.365849 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgtsq\" (UniqueName: \"kubernetes.io/projected/ec562de0-63d3-4613-96f6-b44435bd7dab-kube-api-access-hgtsq\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.424130 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d73ced61-9b8c-475c-b971-362279798b8e-metrics-certs\") pod \"controller-f8648f98b-7c7qd\" (UID: \"d73ced61-9b8c-475c-b971-362279798b8e\") " pod="metallb-system/controller-f8648f98b-7c7qd" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.424282 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d73ced61-9b8c-475c-b971-362279798b8e-cert\") pod \"controller-f8648f98b-7c7qd\" (UID: \"d73ced61-9b8c-475c-b971-362279798b8e\") " pod="metallb-system/controller-f8648f98b-7c7qd" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.424403 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t7vr\" (UniqueName: \"kubernetes.io/projected/d73ced61-9b8c-475c-b971-362279798b8e-kube-api-access-8t7vr\") pod \"controller-f8648f98b-7c7qd\" (UID: \"d73ced61-9b8c-475c-b971-362279798b8e\") " pod="metallb-system/controller-f8648f98b-7c7qd" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.427676 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d73ced61-9b8c-475c-b971-362279798b8e-cert\") pod \"controller-f8648f98b-7c7qd\" (UID: \"d73ced61-9b8c-475c-b971-362279798b8e\") " pod="metallb-system/controller-f8648f98b-7c7qd" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.431976 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d73ced61-9b8c-475c-b971-362279798b8e-metrics-certs\") pod \"controller-f8648f98b-7c7qd\" (UID: \"d73ced61-9b8c-475c-b971-362279798b8e\") " pod="metallb-system/controller-f8648f98b-7c7qd" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.443915 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t7vr\" (UniqueName: \"kubernetes.io/projected/d73ced61-9b8c-475c-b971-362279798b8e-kube-api-access-8t7vr\") pod \"controller-f8648f98b-7c7qd\" (UID: \"d73ced61-9b8c-475c-b971-362279798b8e\") " pod="metallb-system/controller-f8648f98b-7c7qd" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.586376 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-7c7qd" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.627397 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7e1847-fefb-4289-8dbb-c04dec670021-metrics-certs\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.627469 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-l4ld2\" (UID: \"ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.631084 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-l4ld2\" (UID: \"ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.632155 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e7e1847-fefb-4289-8dbb-c04dec670021-metrics-certs\") pod \"frr-k8s-hd27h\" (UID: \"5e7e1847-fefb-4289-8dbb-c04dec670021\") " pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.796780 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-7c7qd"] Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.822545 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.828948 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.829678 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-memberlist\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.829736 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-metrics-certs\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:24 crc kubenswrapper[4806]: E1127 10:33:24.829863 4806 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 27 10:33:24 crc kubenswrapper[4806]: E1127 10:33:24.829945 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-memberlist podName:ec562de0-63d3-4613-96f6-b44435bd7dab nodeName:}" failed. No retries permitted until 2025-11-27 10:33:25.829925119 +0000 UTC m=+710.416516073 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-memberlist") pod "speaker-bvd28" (UID: "ec562de0-63d3-4613-96f6-b44435bd7dab") : secret "metallb-memberlist" not found Nov 27 10:33:24 crc kubenswrapper[4806]: I1127 10:33:24.833741 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-metrics-certs\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:25 crc kubenswrapper[4806]: I1127 10:33:25.104070 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2"] Nov 27 10:33:25 crc kubenswrapper[4806]: W1127 10:33:25.117444 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded9cee58_4b54_4ef2_a86c_99c8f0e83d8b.slice/crio-6eb4828e27fd172fb852e12fbbd83f5ed87017654c77f46354af98b1e775f6ef WatchSource:0}: Error finding container 6eb4828e27fd172fb852e12fbbd83f5ed87017654c77f46354af98b1e775f6ef: Status 404 returned error can't find the container with id 6eb4828e27fd172fb852e12fbbd83f5ed87017654c77f46354af98b1e775f6ef Nov 27 10:33:25 crc kubenswrapper[4806]: I1127 10:33:25.729538 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd27h" event={"ID":"5e7e1847-fefb-4289-8dbb-c04dec670021","Type":"ContainerStarted","Data":"8c8bca12429cf183a2b901bcbb162b583907f464864a4c4cdf1e3938f1630350"} Nov 27 10:33:25 crc kubenswrapper[4806]: I1127 10:33:25.731193 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" event={"ID":"ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b","Type":"ContainerStarted","Data":"6eb4828e27fd172fb852e12fbbd83f5ed87017654c77f46354af98b1e775f6ef"} Nov 27 10:33:25 crc kubenswrapper[4806]: I1127 10:33:25.733045 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-7c7qd" event={"ID":"d73ced61-9b8c-475c-b971-362279798b8e","Type":"ContainerStarted","Data":"81d4eacdc60c40cf55ebe7a39c87c2ee7ffdf67ab0b15f2cccfdf6d3def4bb84"} Nov 27 10:33:25 crc kubenswrapper[4806]: I1127 10:33:25.733147 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-7c7qd" event={"ID":"d73ced61-9b8c-475c-b971-362279798b8e","Type":"ContainerStarted","Data":"00d6664da98277678c78162dbb6088bcfc3b0933d5b20328881dcfdc1e8b354b"} Nov 27 10:33:25 crc kubenswrapper[4806]: I1127 10:33:25.733207 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-7c7qd" event={"ID":"d73ced61-9b8c-475c-b971-362279798b8e","Type":"ContainerStarted","Data":"fb8d1d58972875caf529c4db2a0e425ac2f98b6c1171b889bfb337870e02ef3b"} Nov 27 10:33:25 crc kubenswrapper[4806]: I1127 10:33:25.733623 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-7c7qd" Nov 27 10:33:25 crc kubenswrapper[4806]: I1127 10:33:25.752965 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-7c7qd" podStartSLOduration=1.752937879 podStartE2EDuration="1.752937879s" podCreationTimestamp="2025-11-27 10:33:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:33:25.750671115 +0000 UTC m=+710.337261879" watchObservedRunningTime="2025-11-27 10:33:25.752937879 +0000 UTC m=+710.339528643" Nov 27 10:33:25 crc kubenswrapper[4806]: I1127 10:33:25.843589 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-memberlist\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:25 crc kubenswrapper[4806]: I1127 10:33:25.861274 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ec562de0-63d3-4613-96f6-b44435bd7dab-memberlist\") pod \"speaker-bvd28\" (UID: \"ec562de0-63d3-4613-96f6-b44435bd7dab\") " pod="metallb-system/speaker-bvd28" Nov 27 10:33:26 crc kubenswrapper[4806]: I1127 10:33:26.157323 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bvd28" Nov 27 10:33:26 crc kubenswrapper[4806]: I1127 10:33:26.771139 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bvd28" event={"ID":"ec562de0-63d3-4613-96f6-b44435bd7dab","Type":"ContainerStarted","Data":"cf1efd83ffe3a5a7118867d998b0d3a96acca342430ecfe3c2c313c60d2fdd16"} Nov 27 10:33:26 crc kubenswrapper[4806]: I1127 10:33:26.771525 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bvd28" event={"ID":"ec562de0-63d3-4613-96f6-b44435bd7dab","Type":"ContainerStarted","Data":"5d434857124d5ee7eccf08c0f2c223a1cbb018ac12b9c59bc1f10ab7873df74b"} Nov 27 10:33:27 crc kubenswrapper[4806]: I1127 10:33:27.786551 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bvd28" event={"ID":"ec562de0-63d3-4613-96f6-b44435bd7dab","Type":"ContainerStarted","Data":"1572898859dd10042ebc067217610028809dd3b076f6b5d9c78ffa5ba6125648"} Nov 27 10:33:27 crc kubenswrapper[4806]: I1127 10:33:27.786772 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-bvd28" Nov 27 10:33:34 crc kubenswrapper[4806]: I1127 10:33:34.845643 4806 generic.go:334] "Generic (PLEG): container finished" podID="5e7e1847-fefb-4289-8dbb-c04dec670021" containerID="791df4a4a69497d43d1f31e2ba4568916593abd97af747cfa1e9a24bc46d6f8b" exitCode=0 Nov 27 10:33:34 crc kubenswrapper[4806]: I1127 10:33:34.845736 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd27h" event={"ID":"5e7e1847-fefb-4289-8dbb-c04dec670021","Type":"ContainerDied","Data":"791df4a4a69497d43d1f31e2ba4568916593abd97af747cfa1e9a24bc46d6f8b"} Nov 27 10:33:34 crc kubenswrapper[4806]: I1127 10:33:34.849932 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" event={"ID":"ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b","Type":"ContainerStarted","Data":"d4c9e906f6dea2c21c8d3be107f8e636503f606c319e0aa2ed14f98bf44e0fb9"} Nov 27 10:33:34 crc kubenswrapper[4806]: I1127 10:33:34.882267 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-bvd28" podStartSLOduration=10.882220057 podStartE2EDuration="10.882220057s" podCreationTimestamp="2025-11-27 10:33:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:33:27.821761153 +0000 UTC m=+712.408351917" watchObservedRunningTime="2025-11-27 10:33:34.882220057 +0000 UTC m=+719.468810821" Nov 27 10:33:34 crc kubenswrapper[4806]: I1127 10:33:34.905930 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" podStartSLOduration=3.359996145 podStartE2EDuration="11.90590763s" podCreationTimestamp="2025-11-27 10:33:23 +0000 UTC" firstStartedPulling="2025-11-27 10:33:25.119594149 +0000 UTC m=+709.706184913" lastFinishedPulling="2025-11-27 10:33:33.665505634 +0000 UTC m=+718.252096398" observedRunningTime="2025-11-27 10:33:34.904649855 +0000 UTC m=+719.491240619" watchObservedRunningTime="2025-11-27 10:33:34.90590763 +0000 UTC m=+719.492498394" Nov 27 10:33:35 crc kubenswrapper[4806]: I1127 10:33:35.863413 4806 generic.go:334] "Generic (PLEG): container finished" podID="5e7e1847-fefb-4289-8dbb-c04dec670021" containerID="584b106892852bf7a128e132f50dc71ddc3d085ce16a0fc7dc3e830fb3616153" exitCode=0 Nov 27 10:33:35 crc kubenswrapper[4806]: I1127 10:33:35.863513 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd27h" event={"ID":"5e7e1847-fefb-4289-8dbb-c04dec670021","Type":"ContainerDied","Data":"584b106892852bf7a128e132f50dc71ddc3d085ce16a0fc7dc3e830fb3616153"} Nov 27 10:33:35 crc kubenswrapper[4806]: I1127 10:33:35.864534 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" Nov 27 10:33:36 crc kubenswrapper[4806]: I1127 10:33:36.162879 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-bvd28" Nov 27 10:33:36 crc kubenswrapper[4806]: I1127 10:33:36.871962 4806 generic.go:334] "Generic (PLEG): container finished" podID="5e7e1847-fefb-4289-8dbb-c04dec670021" containerID="7094708977daf4d81ef0fa0539098abce2c99cf175ed6b453a5b0c82206bea0c" exitCode=0 Nov 27 10:33:36 crc kubenswrapper[4806]: I1127 10:33:36.873471 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd27h" event={"ID":"5e7e1847-fefb-4289-8dbb-c04dec670021","Type":"ContainerDied","Data":"7094708977daf4d81ef0fa0539098abce2c99cf175ed6b453a5b0c82206bea0c"} Nov 27 10:33:37 crc kubenswrapper[4806]: I1127 10:33:37.887389 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd27h" event={"ID":"5e7e1847-fefb-4289-8dbb-c04dec670021","Type":"ContainerStarted","Data":"c349621b9b6d99575f6cab4ce0d22609867ceffe8786c13dab9ff814ca4a6d8b"} Nov 27 10:33:37 crc kubenswrapper[4806]: I1127 10:33:37.887779 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd27h" event={"ID":"5e7e1847-fefb-4289-8dbb-c04dec670021","Type":"ContainerStarted","Data":"82ea82fdf47e4adb9249421b51495b4bca9b4a2f611e7d2b1fc6aa35f0aa60ab"} Nov 27 10:33:37 crc kubenswrapper[4806]: I1127 10:33:37.887793 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd27h" event={"ID":"5e7e1847-fefb-4289-8dbb-c04dec670021","Type":"ContainerStarted","Data":"d5505c5deeb1a355b8aac4779fb487465789ee01d7f269c0d92b78ed7519ed08"} Nov 27 10:33:37 crc kubenswrapper[4806]: I1127 10:33:37.887802 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd27h" event={"ID":"5e7e1847-fefb-4289-8dbb-c04dec670021","Type":"ContainerStarted","Data":"bcb9a510a67e93eb2a3b9d876e5a1ac12bf186b26978410c6c3e622a9ad45091"} Nov 27 10:33:37 crc kubenswrapper[4806]: I1127 10:33:37.887810 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd27h" event={"ID":"5e7e1847-fefb-4289-8dbb-c04dec670021","Type":"ContainerStarted","Data":"2340b6509110d09278de0340e59f9b88a0a60c56d76f137dee29f467a0f74a79"} Nov 27 10:33:38 crc kubenswrapper[4806]: I1127 10:33:38.901130 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd27h" event={"ID":"5e7e1847-fefb-4289-8dbb-c04dec670021","Type":"ContainerStarted","Data":"28b72886af14bace7586572c2d56584d5cbfb55049da6e423f3434ae6af9d79a"} Nov 27 10:33:38 crc kubenswrapper[4806]: I1127 10:33:38.902022 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:38 crc kubenswrapper[4806]: I1127 10:33:38.945904 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-hd27h" podStartSLOduration=7.261983927 podStartE2EDuration="15.945860844s" podCreationTimestamp="2025-11-27 10:33:23 +0000 UTC" firstStartedPulling="2025-11-27 10:33:25.005069193 +0000 UTC m=+709.591659957" lastFinishedPulling="2025-11-27 10:33:33.68894611 +0000 UTC m=+718.275536874" observedRunningTime="2025-11-27 10:33:38.943629011 +0000 UTC m=+723.530219805" watchObservedRunningTime="2025-11-27 10:33:38.945860844 +0000 UTC m=+723.532451608" Nov 27 10:33:39 crc kubenswrapper[4806]: I1127 10:33:39.541626 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-x2bd9"] Nov 27 10:33:39 crc kubenswrapper[4806]: I1127 10:33:39.542585 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-x2bd9" Nov 27 10:33:39 crc kubenswrapper[4806]: I1127 10:33:39.545397 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-jgp42" Nov 27 10:33:39 crc kubenswrapper[4806]: I1127 10:33:39.546253 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 27 10:33:39 crc kubenswrapper[4806]: I1127 10:33:39.547708 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 27 10:33:39 crc kubenswrapper[4806]: I1127 10:33:39.608283 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-x2bd9"] Nov 27 10:33:39 crc kubenswrapper[4806]: I1127 10:33:39.682265 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4ppp\" (UniqueName: \"kubernetes.io/projected/3446ae4d-990b-409c-a5d4-ad0be0749d32-kube-api-access-s4ppp\") pod \"openstack-operator-index-x2bd9\" (UID: \"3446ae4d-990b-409c-a5d4-ad0be0749d32\") " pod="openstack-operators/openstack-operator-index-x2bd9" Nov 27 10:33:39 crc kubenswrapper[4806]: I1127 10:33:39.784591 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4ppp\" (UniqueName: \"kubernetes.io/projected/3446ae4d-990b-409c-a5d4-ad0be0749d32-kube-api-access-s4ppp\") pod \"openstack-operator-index-x2bd9\" (UID: \"3446ae4d-990b-409c-a5d4-ad0be0749d32\") " pod="openstack-operators/openstack-operator-index-x2bd9" Nov 27 10:33:39 crc kubenswrapper[4806]: I1127 10:33:39.807800 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4ppp\" (UniqueName: \"kubernetes.io/projected/3446ae4d-990b-409c-a5d4-ad0be0749d32-kube-api-access-s4ppp\") pod \"openstack-operator-index-x2bd9\" (UID: \"3446ae4d-990b-409c-a5d4-ad0be0749d32\") " pod="openstack-operators/openstack-operator-index-x2bd9" Nov 27 10:33:39 crc kubenswrapper[4806]: I1127 10:33:39.823824 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:39 crc kubenswrapper[4806]: I1127 10:33:39.859727 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-x2bd9" Nov 27 10:33:39 crc kubenswrapper[4806]: I1127 10:33:39.915946 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-hd27h" Nov 27 10:33:40 crc kubenswrapper[4806]: I1127 10:33:40.130496 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-x2bd9"] Nov 27 10:33:40 crc kubenswrapper[4806]: I1127 10:33:40.919897 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-x2bd9" event={"ID":"3446ae4d-990b-409c-a5d4-ad0be0749d32","Type":"ContainerStarted","Data":"2c88e774eab5d840a80fa6f50b7893b9e23b22d5cb103bd2185ac2f810eb013f"} Nov 27 10:33:42 crc kubenswrapper[4806]: I1127 10:33:42.916747 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-x2bd9"] Nov 27 10:33:42 crc kubenswrapper[4806]: I1127 10:33:42.936174 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-x2bd9" event={"ID":"3446ae4d-990b-409c-a5d4-ad0be0749d32","Type":"ContainerStarted","Data":"69769ea053aa8288f89f18fc4a136c5997d6aefa7af4173036a9b094dcd848ce"} Nov 27 10:33:42 crc kubenswrapper[4806]: I1127 10:33:42.966068 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-x2bd9" podStartSLOduration=1.5606408109999999 podStartE2EDuration="3.966046875s" podCreationTimestamp="2025-11-27 10:33:39 +0000 UTC" firstStartedPulling="2025-11-27 10:33:40.139874822 +0000 UTC m=+724.726465586" lastFinishedPulling="2025-11-27 10:33:42.545280886 +0000 UTC m=+727.131871650" observedRunningTime="2025-11-27 10:33:42.965367326 +0000 UTC m=+727.551958130" watchObservedRunningTime="2025-11-27 10:33:42.966046875 +0000 UTC m=+727.552637639" Nov 27 10:33:43 crc kubenswrapper[4806]: I1127 10:33:43.528363 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-qjct2"] Nov 27 10:33:43 crc kubenswrapper[4806]: I1127 10:33:43.531439 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qjct2" Nov 27 10:33:43 crc kubenswrapper[4806]: I1127 10:33:43.558601 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qjct2"] Nov 27 10:33:43 crc kubenswrapper[4806]: I1127 10:33:43.642171 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkclw\" (UniqueName: \"kubernetes.io/projected/7a4448af-a960-4299-8ed2-71b93c8b7e46-kube-api-access-bkclw\") pod \"openstack-operator-index-qjct2\" (UID: \"7a4448af-a960-4299-8ed2-71b93c8b7e46\") " pod="openstack-operators/openstack-operator-index-qjct2" Nov 27 10:33:43 crc kubenswrapper[4806]: I1127 10:33:43.744115 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkclw\" (UniqueName: \"kubernetes.io/projected/7a4448af-a960-4299-8ed2-71b93c8b7e46-kube-api-access-bkclw\") pod \"openstack-operator-index-qjct2\" (UID: \"7a4448af-a960-4299-8ed2-71b93c8b7e46\") " pod="openstack-operators/openstack-operator-index-qjct2" Nov 27 10:33:43 crc kubenswrapper[4806]: I1127 10:33:43.771600 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkclw\" (UniqueName: \"kubernetes.io/projected/7a4448af-a960-4299-8ed2-71b93c8b7e46-kube-api-access-bkclw\") pod \"openstack-operator-index-qjct2\" (UID: \"7a4448af-a960-4299-8ed2-71b93c8b7e46\") " pod="openstack-operators/openstack-operator-index-qjct2" Nov 27 10:33:43 crc kubenswrapper[4806]: I1127 10:33:43.863543 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qjct2" Nov 27 10:33:43 crc kubenswrapper[4806]: I1127 10:33:43.946025 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-x2bd9" podUID="3446ae4d-990b-409c-a5d4-ad0be0749d32" containerName="registry-server" containerID="cri-o://69769ea053aa8288f89f18fc4a136c5997d6aefa7af4173036a9b094dcd848ce" gracePeriod=2 Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.093435 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qjct2"] Nov 27 10:33:44 crc kubenswrapper[4806]: W1127 10:33:44.100360 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a4448af_a960_4299_8ed2_71b93c8b7e46.slice/crio-2789ad3d22e707daba479d945e91ac759f41a6977bd93fc6451e6f37604ec2cc WatchSource:0}: Error finding container 2789ad3d22e707daba479d945e91ac759f41a6977bd93fc6451e6f37604ec2cc: Status 404 returned error can't find the container with id 2789ad3d22e707daba479d945e91ac759f41a6977bd93fc6451e6f37604ec2cc Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.388017 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-x2bd9" Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.450260 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.450314 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.464438 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4ppp\" (UniqueName: \"kubernetes.io/projected/3446ae4d-990b-409c-a5d4-ad0be0749d32-kube-api-access-s4ppp\") pod \"3446ae4d-990b-409c-a5d4-ad0be0749d32\" (UID: \"3446ae4d-990b-409c-a5d4-ad0be0749d32\") " Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.469404 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3446ae4d-990b-409c-a5d4-ad0be0749d32-kube-api-access-s4ppp" (OuterVolumeSpecName: "kube-api-access-s4ppp") pod "3446ae4d-990b-409c-a5d4-ad0be0749d32" (UID: "3446ae4d-990b-409c-a5d4-ad0be0749d32"). InnerVolumeSpecName "kube-api-access-s4ppp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.566020 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4ppp\" (UniqueName: \"kubernetes.io/projected/3446ae4d-990b-409c-a5d4-ad0be0749d32-kube-api-access-s4ppp\") on node \"crc\" DevicePath \"\"" Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.590528 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-7c7qd" Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.834777 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-l4ld2" Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.956174 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qjct2" event={"ID":"7a4448af-a960-4299-8ed2-71b93c8b7e46","Type":"ContainerStarted","Data":"c7a87f1e3d226c696576e6a76481c3f3e4c173b95b942c67a9c05b310eaf0fe8"} Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.956255 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qjct2" event={"ID":"7a4448af-a960-4299-8ed2-71b93c8b7e46","Type":"ContainerStarted","Data":"2789ad3d22e707daba479d945e91ac759f41a6977bd93fc6451e6f37604ec2cc"} Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.957336 4806 generic.go:334] "Generic (PLEG): container finished" podID="3446ae4d-990b-409c-a5d4-ad0be0749d32" containerID="69769ea053aa8288f89f18fc4a136c5997d6aefa7af4173036a9b094dcd848ce" exitCode=0 Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.957375 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-x2bd9" event={"ID":"3446ae4d-990b-409c-a5d4-ad0be0749d32","Type":"ContainerDied","Data":"69769ea053aa8288f89f18fc4a136c5997d6aefa7af4173036a9b094dcd848ce"} Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.957394 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-x2bd9" Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.957409 4806 scope.go:117] "RemoveContainer" containerID="69769ea053aa8288f89f18fc4a136c5997d6aefa7af4173036a9b094dcd848ce" Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.957398 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-x2bd9" event={"ID":"3446ae4d-990b-409c-a5d4-ad0be0749d32","Type":"ContainerDied","Data":"2c88e774eab5d840a80fa6f50b7893b9e23b22d5cb103bd2185ac2f810eb013f"} Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.972248 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-qjct2" podStartSLOduration=1.909022105 podStartE2EDuration="1.972215551s" podCreationTimestamp="2025-11-27 10:33:43 +0000 UTC" firstStartedPulling="2025-11-27 10:33:44.105918181 +0000 UTC m=+728.692508945" lastFinishedPulling="2025-11-27 10:33:44.169111627 +0000 UTC m=+728.755702391" observedRunningTime="2025-11-27 10:33:44.969961838 +0000 UTC m=+729.556552602" watchObservedRunningTime="2025-11-27 10:33:44.972215551 +0000 UTC m=+729.558806315" Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.986450 4806 scope.go:117] "RemoveContainer" containerID="69769ea053aa8288f89f18fc4a136c5997d6aefa7af4173036a9b094dcd848ce" Nov 27 10:33:44 crc kubenswrapper[4806]: E1127 10:33:44.987118 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69769ea053aa8288f89f18fc4a136c5997d6aefa7af4173036a9b094dcd848ce\": container with ID starting with 69769ea053aa8288f89f18fc4a136c5997d6aefa7af4173036a9b094dcd848ce not found: ID does not exist" containerID="69769ea053aa8288f89f18fc4a136c5997d6aefa7af4173036a9b094dcd848ce" Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.987159 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69769ea053aa8288f89f18fc4a136c5997d6aefa7af4173036a9b094dcd848ce"} err="failed to get container status \"69769ea053aa8288f89f18fc4a136c5997d6aefa7af4173036a9b094dcd848ce\": rpc error: code = NotFound desc = could not find container \"69769ea053aa8288f89f18fc4a136c5997d6aefa7af4173036a9b094dcd848ce\": container with ID starting with 69769ea053aa8288f89f18fc4a136c5997d6aefa7af4173036a9b094dcd848ce not found: ID does not exist" Nov 27 10:33:44 crc kubenswrapper[4806]: I1127 10:33:44.997096 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-x2bd9"] Nov 27 10:33:45 crc kubenswrapper[4806]: I1127 10:33:45.001138 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-x2bd9"] Nov 27 10:33:46 crc kubenswrapper[4806]: I1127 10:33:46.131047 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3446ae4d-990b-409c-a5d4-ad0be0749d32" path="/var/lib/kubelet/pods/3446ae4d-990b-409c-a5d4-ad0be0749d32/volumes" Nov 27 10:33:53 crc kubenswrapper[4806]: I1127 10:33:53.864417 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-qjct2" Nov 27 10:33:53 crc kubenswrapper[4806]: I1127 10:33:53.864958 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-qjct2" Nov 27 10:33:53 crc kubenswrapper[4806]: I1127 10:33:53.894267 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-qjct2" Nov 27 10:33:54 crc kubenswrapper[4806]: I1127 10:33:54.052896 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-qjct2" Nov 27 10:33:54 crc kubenswrapper[4806]: I1127 10:33:54.829843 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-hd27h" Nov 27 10:34:01 crc kubenswrapper[4806]: I1127 10:34:01.924477 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq"] Nov 27 10:34:01 crc kubenswrapper[4806]: E1127 10:34:01.926177 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3446ae4d-990b-409c-a5d4-ad0be0749d32" containerName="registry-server" Nov 27 10:34:01 crc kubenswrapper[4806]: I1127 10:34:01.926275 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3446ae4d-990b-409c-a5d4-ad0be0749d32" containerName="registry-server" Nov 27 10:34:01 crc kubenswrapper[4806]: I1127 10:34:01.926461 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3446ae4d-990b-409c-a5d4-ad0be0749d32" containerName="registry-server" Nov 27 10:34:01 crc kubenswrapper[4806]: I1127 10:34:01.927298 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" Nov 27 10:34:01 crc kubenswrapper[4806]: I1127 10:34:01.929745 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-g7trx" Nov 27 10:34:01 crc kubenswrapper[4806]: I1127 10:34:01.948503 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq"] Nov 27 10:34:01 crc kubenswrapper[4806]: I1127 10:34:01.950127 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg7z9\" (UniqueName: \"kubernetes.io/projected/99394f5f-eef5-416a-ae48-376b3e6f6122-kube-api-access-jg7z9\") pod \"d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq\" (UID: \"99394f5f-eef5-416a-ae48-376b3e6f6122\") " pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" Nov 27 10:34:01 crc kubenswrapper[4806]: I1127 10:34:01.950181 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/99394f5f-eef5-416a-ae48-376b3e6f6122-bundle\") pod \"d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq\" (UID: \"99394f5f-eef5-416a-ae48-376b3e6f6122\") " pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" Nov 27 10:34:01 crc kubenswrapper[4806]: I1127 10:34:01.950213 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/99394f5f-eef5-416a-ae48-376b3e6f6122-util\") pod \"d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq\" (UID: \"99394f5f-eef5-416a-ae48-376b3e6f6122\") " pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" Nov 27 10:34:02 crc kubenswrapper[4806]: I1127 10:34:02.051810 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/99394f5f-eef5-416a-ae48-376b3e6f6122-bundle\") pod \"d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq\" (UID: \"99394f5f-eef5-416a-ae48-376b3e6f6122\") " pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" Nov 27 10:34:02 crc kubenswrapper[4806]: I1127 10:34:02.051884 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/99394f5f-eef5-416a-ae48-376b3e6f6122-util\") pod \"d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq\" (UID: \"99394f5f-eef5-416a-ae48-376b3e6f6122\") " pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" Nov 27 10:34:02 crc kubenswrapper[4806]: I1127 10:34:02.052016 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg7z9\" (UniqueName: \"kubernetes.io/projected/99394f5f-eef5-416a-ae48-376b3e6f6122-kube-api-access-jg7z9\") pod \"d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq\" (UID: \"99394f5f-eef5-416a-ae48-376b3e6f6122\") " pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" Nov 27 10:34:02 crc kubenswrapper[4806]: I1127 10:34:02.052590 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/99394f5f-eef5-416a-ae48-376b3e6f6122-bundle\") pod \"d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq\" (UID: \"99394f5f-eef5-416a-ae48-376b3e6f6122\") " pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" Nov 27 10:34:02 crc kubenswrapper[4806]: I1127 10:34:02.052633 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/99394f5f-eef5-416a-ae48-376b3e6f6122-util\") pod \"d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq\" (UID: \"99394f5f-eef5-416a-ae48-376b3e6f6122\") " pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" Nov 27 10:34:02 crc kubenswrapper[4806]: I1127 10:34:02.077588 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg7z9\" (UniqueName: \"kubernetes.io/projected/99394f5f-eef5-416a-ae48-376b3e6f6122-kube-api-access-jg7z9\") pod \"d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq\" (UID: \"99394f5f-eef5-416a-ae48-376b3e6f6122\") " pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" Nov 27 10:34:02 crc kubenswrapper[4806]: I1127 10:34:02.255861 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" Nov 27 10:34:02 crc kubenswrapper[4806]: I1127 10:34:02.791724 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq"] Nov 27 10:34:02 crc kubenswrapper[4806]: W1127 10:34:02.796587 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99394f5f_eef5_416a_ae48_376b3e6f6122.slice/crio-d2408c594737390a0a0c48a14c6bbf31632028cf4c95d438c7b88e376476c159 WatchSource:0}: Error finding container d2408c594737390a0a0c48a14c6bbf31632028cf4c95d438c7b88e376476c159: Status 404 returned error can't find the container with id d2408c594737390a0a0c48a14c6bbf31632028cf4c95d438c7b88e376476c159 Nov 27 10:34:03 crc kubenswrapper[4806]: I1127 10:34:03.096382 4806 generic.go:334] "Generic (PLEG): container finished" podID="99394f5f-eef5-416a-ae48-376b3e6f6122" containerID="d04638a3b76fd81d96d98c8e1b99e28c7d90b143455d63a52cc8c2a22ce6701b" exitCode=0 Nov 27 10:34:03 crc kubenswrapper[4806]: I1127 10:34:03.096431 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" event={"ID":"99394f5f-eef5-416a-ae48-376b3e6f6122","Type":"ContainerDied","Data":"d04638a3b76fd81d96d98c8e1b99e28c7d90b143455d63a52cc8c2a22ce6701b"} Nov 27 10:34:03 crc kubenswrapper[4806]: I1127 10:34:03.096464 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" event={"ID":"99394f5f-eef5-416a-ae48-376b3e6f6122","Type":"ContainerStarted","Data":"d2408c594737390a0a0c48a14c6bbf31632028cf4c95d438c7b88e376476c159"} Nov 27 10:34:04 crc kubenswrapper[4806]: I1127 10:34:04.109590 4806 generic.go:334] "Generic (PLEG): container finished" podID="99394f5f-eef5-416a-ae48-376b3e6f6122" containerID="1b06766a668eb9e613edd26a4b867edef9fd79415a223c519e46141aa5aa71ec" exitCode=0 Nov 27 10:34:04 crc kubenswrapper[4806]: I1127 10:34:04.109658 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" event={"ID":"99394f5f-eef5-416a-ae48-376b3e6f6122","Type":"ContainerDied","Data":"1b06766a668eb9e613edd26a4b867edef9fd79415a223c519e46141aa5aa71ec"} Nov 27 10:34:05 crc kubenswrapper[4806]: I1127 10:34:05.127886 4806 generic.go:334] "Generic (PLEG): container finished" podID="99394f5f-eef5-416a-ae48-376b3e6f6122" containerID="d21f1c83450950dde1d1c995d68f41060d78062d6b9a1a817760771e28589e91" exitCode=0 Nov 27 10:34:05 crc kubenswrapper[4806]: I1127 10:34:05.128017 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" event={"ID":"99394f5f-eef5-416a-ae48-376b3e6f6122","Type":"ContainerDied","Data":"d21f1c83450950dde1d1c995d68f41060d78062d6b9a1a817760771e28589e91"} Nov 27 10:34:06 crc kubenswrapper[4806]: I1127 10:34:06.407154 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" Nov 27 10:34:06 crc kubenswrapper[4806]: I1127 10:34:06.521386 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/99394f5f-eef5-416a-ae48-376b3e6f6122-bundle\") pod \"99394f5f-eef5-416a-ae48-376b3e6f6122\" (UID: \"99394f5f-eef5-416a-ae48-376b3e6f6122\") " Nov 27 10:34:06 crc kubenswrapper[4806]: I1127 10:34:06.521490 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/99394f5f-eef5-416a-ae48-376b3e6f6122-util\") pod \"99394f5f-eef5-416a-ae48-376b3e6f6122\" (UID: \"99394f5f-eef5-416a-ae48-376b3e6f6122\") " Nov 27 10:34:06 crc kubenswrapper[4806]: I1127 10:34:06.521587 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg7z9\" (UniqueName: \"kubernetes.io/projected/99394f5f-eef5-416a-ae48-376b3e6f6122-kube-api-access-jg7z9\") pod \"99394f5f-eef5-416a-ae48-376b3e6f6122\" (UID: \"99394f5f-eef5-416a-ae48-376b3e6f6122\") " Nov 27 10:34:06 crc kubenswrapper[4806]: I1127 10:34:06.525552 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99394f5f-eef5-416a-ae48-376b3e6f6122-bundle" (OuterVolumeSpecName: "bundle") pod "99394f5f-eef5-416a-ae48-376b3e6f6122" (UID: "99394f5f-eef5-416a-ae48-376b3e6f6122"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:34:06 crc kubenswrapper[4806]: I1127 10:34:06.535833 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99394f5f-eef5-416a-ae48-376b3e6f6122-util" (OuterVolumeSpecName: "util") pod "99394f5f-eef5-416a-ae48-376b3e6f6122" (UID: "99394f5f-eef5-416a-ae48-376b3e6f6122"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:34:06 crc kubenswrapper[4806]: I1127 10:34:06.536511 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99394f5f-eef5-416a-ae48-376b3e6f6122-kube-api-access-jg7z9" (OuterVolumeSpecName: "kube-api-access-jg7z9") pod "99394f5f-eef5-416a-ae48-376b3e6f6122" (UID: "99394f5f-eef5-416a-ae48-376b3e6f6122"). InnerVolumeSpecName "kube-api-access-jg7z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:34:06 crc kubenswrapper[4806]: I1127 10:34:06.626392 4806 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/99394f5f-eef5-416a-ae48-376b3e6f6122-util\") on node \"crc\" DevicePath \"\"" Nov 27 10:34:06 crc kubenswrapper[4806]: I1127 10:34:06.626445 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg7z9\" (UniqueName: \"kubernetes.io/projected/99394f5f-eef5-416a-ae48-376b3e6f6122-kube-api-access-jg7z9\") on node \"crc\" DevicePath \"\"" Nov 27 10:34:06 crc kubenswrapper[4806]: I1127 10:34:06.626462 4806 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/99394f5f-eef5-416a-ae48-376b3e6f6122-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:34:07 crc kubenswrapper[4806]: I1127 10:34:07.151331 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" event={"ID":"99394f5f-eef5-416a-ae48-376b3e6f6122","Type":"ContainerDied","Data":"d2408c594737390a0a0c48a14c6bbf31632028cf4c95d438c7b88e376476c159"} Nov 27 10:34:07 crc kubenswrapper[4806]: I1127 10:34:07.151430 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2408c594737390a0a0c48a14c6bbf31632028cf4c95d438c7b88e376476c159" Nov 27 10:34:07 crc kubenswrapper[4806]: I1127 10:34:07.151431 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.130398 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6784745db8-rng9q"] Nov 27 10:34:14 crc kubenswrapper[4806]: E1127 10:34:14.131376 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99394f5f-eef5-416a-ae48-376b3e6f6122" containerName="extract" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.131391 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="99394f5f-eef5-416a-ae48-376b3e6f6122" containerName="extract" Nov 27 10:34:14 crc kubenswrapper[4806]: E1127 10:34:14.131411 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99394f5f-eef5-416a-ae48-376b3e6f6122" containerName="util" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.131417 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="99394f5f-eef5-416a-ae48-376b3e6f6122" containerName="util" Nov 27 10:34:14 crc kubenswrapper[4806]: E1127 10:34:14.131435 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99394f5f-eef5-416a-ae48-376b3e6f6122" containerName="pull" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.131442 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="99394f5f-eef5-416a-ae48-376b3e6f6122" containerName="pull" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.131568 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="99394f5f-eef5-416a-ae48-376b3e6f6122" containerName="extract" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.132151 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6784745db8-rng9q" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.135724 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-h9blx" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.171960 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6784745db8-rng9q"] Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.242973 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgrh5\" (UniqueName: \"kubernetes.io/projected/33c3ee1d-c0cc-4c75-b6e9-0059dd8042c7-kube-api-access-wgrh5\") pod \"openstack-operator-controller-operator-6784745db8-rng9q\" (UID: \"33c3ee1d-c0cc-4c75-b6e9-0059dd8042c7\") " pod="openstack-operators/openstack-operator-controller-operator-6784745db8-rng9q" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.345508 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgrh5\" (UniqueName: \"kubernetes.io/projected/33c3ee1d-c0cc-4c75-b6e9-0059dd8042c7-kube-api-access-wgrh5\") pod \"openstack-operator-controller-operator-6784745db8-rng9q\" (UID: \"33c3ee1d-c0cc-4c75-b6e9-0059dd8042c7\") " pod="openstack-operators/openstack-operator-controller-operator-6784745db8-rng9q" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.366937 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgrh5\" (UniqueName: \"kubernetes.io/projected/33c3ee1d-c0cc-4c75-b6e9-0059dd8042c7-kube-api-access-wgrh5\") pod \"openstack-operator-controller-operator-6784745db8-rng9q\" (UID: \"33c3ee1d-c0cc-4c75-b6e9-0059dd8042c7\") " pod="openstack-operators/openstack-operator-controller-operator-6784745db8-rng9q" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.449756 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6784745db8-rng9q" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.454397 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.454492 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.454598 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.455366 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5e15c67a8ce5b7fd1f5bca346e1724cf329b9d00aa70feed8824c46529a879e6"} pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.455439 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" containerID="cri-o://5e15c67a8ce5b7fd1f5bca346e1724cf329b9d00aa70feed8824c46529a879e6" gracePeriod=600 Nov 27 10:34:14 crc kubenswrapper[4806]: I1127 10:34:14.943332 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6784745db8-rng9q"] Nov 27 10:34:14 crc kubenswrapper[4806]: W1127 10:34:14.956214 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33c3ee1d_c0cc_4c75_b6e9_0059dd8042c7.slice/crio-ce219167b306e2800b857df1a4ed82c9772f1674b50698bbb0be779527091c15 WatchSource:0}: Error finding container ce219167b306e2800b857df1a4ed82c9772f1674b50698bbb0be779527091c15: Status 404 returned error can't find the container with id ce219167b306e2800b857df1a4ed82c9772f1674b50698bbb0be779527091c15 Nov 27 10:34:15 crc kubenswrapper[4806]: I1127 10:34:15.228488 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6784745db8-rng9q" event={"ID":"33c3ee1d-c0cc-4c75-b6e9-0059dd8042c7","Type":"ContainerStarted","Data":"ce219167b306e2800b857df1a4ed82c9772f1674b50698bbb0be779527091c15"} Nov 27 10:34:15 crc kubenswrapper[4806]: I1127 10:34:15.232703 4806 generic.go:334] "Generic (PLEG): container finished" podID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerID="5e15c67a8ce5b7fd1f5bca346e1724cf329b9d00aa70feed8824c46529a879e6" exitCode=0 Nov 27 10:34:15 crc kubenswrapper[4806]: I1127 10:34:15.232755 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerDied","Data":"5e15c67a8ce5b7fd1f5bca346e1724cf329b9d00aa70feed8824c46529a879e6"} Nov 27 10:34:15 crc kubenswrapper[4806]: I1127 10:34:15.232786 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"dc655ac750298c594d4faac6cef9782fc2d5a20661ea69277b9fafe78a2e81f0"} Nov 27 10:34:15 crc kubenswrapper[4806]: I1127 10:34:15.232806 4806 scope.go:117] "RemoveContainer" containerID="27aa6f365485f368ff18494a5978d93bb6e961422bea73da5d7f2d1830e740c0" Nov 27 10:34:21 crc kubenswrapper[4806]: I1127 10:34:21.282638 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6784745db8-rng9q" event={"ID":"33c3ee1d-c0cc-4c75-b6e9-0059dd8042c7","Type":"ContainerStarted","Data":"9ef78ce97633e140c0ceafc9a485ede91c05a359a4a607731949bba129b5cb0c"} Nov 27 10:34:21 crc kubenswrapper[4806]: I1127 10:34:21.284217 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6784745db8-rng9q" Nov 27 10:34:21 crc kubenswrapper[4806]: I1127 10:34:21.316213 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6784745db8-rng9q" podStartSLOduration=2.018915729 podStartE2EDuration="7.316196971s" podCreationTimestamp="2025-11-27 10:34:14 +0000 UTC" firstStartedPulling="2025-11-27 10:34:14.959073319 +0000 UTC m=+759.545664083" lastFinishedPulling="2025-11-27 10:34:20.256354561 +0000 UTC m=+764.842945325" observedRunningTime="2025-11-27 10:34:21.312706643 +0000 UTC m=+765.899297407" watchObservedRunningTime="2025-11-27 10:34:21.316196971 +0000 UTC m=+765.902787735" Nov 27 10:34:34 crc kubenswrapper[4806]: I1127 10:34:34.453441 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6784745db8-rng9q" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.387811 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.390066 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.403284 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-vm94m" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.408714 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.420670 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.422318 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.428372 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-5jczx" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.458923 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.469157 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-ct9ps"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.476950 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-ct9ps" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.491473 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-tgklg" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.493525 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmzcm\" (UniqueName: \"kubernetes.io/projected/96cf5347-61b7-4ecb-8093-59b25482cd54-kube-api-access-mmzcm\") pod \"barbican-operator-controller-manager-7b64f4fb85-kn7j2\" (UID: \"96cf5347-61b7-4ecb-8093-59b25482cd54\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.500620 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-ct9ps"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.512466 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.513725 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.517296 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-kgthn" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.554675 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.566782 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.568045 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.572559 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-4zfj8" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.579313 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.581051 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.591821 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-ktdqp" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.594670 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgwj9\" (UniqueName: \"kubernetes.io/projected/33c0104b-4a8a-46fa-bd83-b42f86f96789-kube-api-access-fgwj9\") pod \"cinder-operator-controller-manager-6b7f75547b-lcsw6\" (UID: \"33c0104b-4a8a-46fa-bd83-b42f86f96789\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.594761 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8wht\" (UniqueName: \"kubernetes.io/projected/93cca2e7-6982-42d3-8848-10e4f7a70509-kube-api-access-n8wht\") pod \"designate-operator-controller-manager-955677c94-ct9ps\" (UID: \"93cca2e7-6982-42d3-8848-10e4f7a70509\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-ct9ps" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.594799 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmzcm\" (UniqueName: \"kubernetes.io/projected/96cf5347-61b7-4ecb-8093-59b25482cd54-kube-api-access-mmzcm\") pod \"barbican-operator-controller-manager-7b64f4fb85-kn7j2\" (UID: \"96cf5347-61b7-4ecb-8093-59b25482cd54\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.603941 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.656387 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.696746 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54v62\" (UniqueName: \"kubernetes.io/projected/5b227bdd-2c69-4053-a116-8085cd157a93-kube-api-access-54v62\") pod \"horizon-operator-controller-manager-5d494799bf-v8r26\" (UID: \"5b227bdd-2c69-4053-a116-8085cd157a93\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.707470 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xngs\" (UniqueName: \"kubernetes.io/projected/94f64657-d78c-4f52-8aa1-b23b61d5af68-kube-api-access-7xngs\") pod \"heat-operator-controller-manager-5b77f656f-czdn8\" (UID: \"94f64657-d78c-4f52-8aa1-b23b61d5af68\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.707817 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgwj9\" (UniqueName: \"kubernetes.io/projected/33c0104b-4a8a-46fa-bd83-b42f86f96789-kube-api-access-fgwj9\") pod \"cinder-operator-controller-manager-6b7f75547b-lcsw6\" (UID: \"33c0104b-4a8a-46fa-bd83-b42f86f96789\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.708086 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m6zb\" (UniqueName: \"kubernetes.io/projected/b6163410-3453-43c3-8cd7-5a354e447025-kube-api-access-2m6zb\") pod \"glance-operator-controller-manager-589cbd6b5b-9jh67\" (UID: \"b6163410-3453-43c3-8cd7-5a354e447025\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.708202 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8wht\" (UniqueName: \"kubernetes.io/projected/93cca2e7-6982-42d3-8848-10e4f7a70509-kube-api-access-n8wht\") pod \"designate-operator-controller-manager-955677c94-ct9ps\" (UID: \"93cca2e7-6982-42d3-8848-10e4f7a70509\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-ct9ps" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.710044 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmzcm\" (UniqueName: \"kubernetes.io/projected/96cf5347-61b7-4ecb-8093-59b25482cd54-kube-api-access-mmzcm\") pod \"barbican-operator-controller-manager-7b64f4fb85-kn7j2\" (UID: \"96cf5347-61b7-4ecb-8093-59b25482cd54\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.738435 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.739291 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-55nb9"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.750690 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.772565 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgwj9\" (UniqueName: \"kubernetes.io/projected/33c0104b-4a8a-46fa-bd83-b42f86f96789-kube-api-access-fgwj9\") pod \"cinder-operator-controller-manager-6b7f75547b-lcsw6\" (UID: \"33c0104b-4a8a-46fa-bd83-b42f86f96789\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.783821 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.784635 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-dgf8z" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.809165 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8wht\" (UniqueName: \"kubernetes.io/projected/93cca2e7-6982-42d3-8848-10e4f7a70509-kube-api-access-n8wht\") pod \"designate-operator-controller-manager-955677c94-ct9ps\" (UID: \"93cca2e7-6982-42d3-8848-10e4f7a70509\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-ct9ps" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.809923 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m6zb\" (UniqueName: \"kubernetes.io/projected/b6163410-3453-43c3-8cd7-5a354e447025-kube-api-access-2m6zb\") pod \"glance-operator-controller-manager-589cbd6b5b-9jh67\" (UID: \"b6163410-3453-43c3-8cd7-5a354e447025\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.809986 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54v62\" (UniqueName: \"kubernetes.io/projected/5b227bdd-2c69-4053-a116-8085cd157a93-kube-api-access-54v62\") pod \"horizon-operator-controller-manager-5d494799bf-v8r26\" (UID: \"5b227bdd-2c69-4053-a116-8085cd157a93\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.810035 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88wlx\" (UniqueName: \"kubernetes.io/projected/60db26f8-5bc6-477c-bf21-6842e1b3a9db-kube-api-access-88wlx\") pod \"infra-operator-controller-manager-57548d458d-55nb9\" (UID: \"60db26f8-5bc6-477c-bf21-6842e1b3a9db\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.810075 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert\") pod \"infra-operator-controller-manager-57548d458d-55nb9\" (UID: \"60db26f8-5bc6-477c-bf21-6842e1b3a9db\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.810103 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xngs\" (UniqueName: \"kubernetes.io/projected/94f64657-d78c-4f52-8aa1-b23b61d5af68-kube-api-access-7xngs\") pod \"heat-operator-controller-manager-5b77f656f-czdn8\" (UID: \"94f64657-d78c-4f52-8aa1-b23b61d5af68\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.820121 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-ct9ps" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.843876 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xngs\" (UniqueName: \"kubernetes.io/projected/94f64657-d78c-4f52-8aa1-b23b61d5af68-kube-api-access-7xngs\") pod \"heat-operator-controller-manager-5b77f656f-czdn8\" (UID: \"94f64657-d78c-4f52-8aa1-b23b61d5af68\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.852388 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-55nb9"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.871786 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54v62\" (UniqueName: \"kubernetes.io/projected/5b227bdd-2c69-4053-a116-8085cd157a93-kube-api-access-54v62\") pod \"horizon-operator-controller-manager-5d494799bf-v8r26\" (UID: \"5b227bdd-2c69-4053-a116-8085cd157a93\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.872102 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.874147 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.884874 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-5mv25" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.886560 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m6zb\" (UniqueName: \"kubernetes.io/projected/b6163410-3453-43c3-8cd7-5a354e447025-kube-api-access-2m6zb\") pod \"glance-operator-controller-manager-589cbd6b5b-9jh67\" (UID: \"b6163410-3453-43c3-8cd7-5a354e447025\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.902993 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.914159 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88wlx\" (UniqueName: \"kubernetes.io/projected/60db26f8-5bc6-477c-bf21-6842e1b3a9db-kube-api-access-88wlx\") pod \"infra-operator-controller-manager-57548d458d-55nb9\" (UID: \"60db26f8-5bc6-477c-bf21-6842e1b3a9db\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.914221 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert\") pod \"infra-operator-controller-manager-57548d458d-55nb9\" (UID: \"60db26f8-5bc6-477c-bf21-6842e1b3a9db\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:01 crc kubenswrapper[4806]: E1127 10:35:01.914392 4806 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 10:35:01 crc kubenswrapper[4806]: E1127 10:35:01.914450 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert podName:60db26f8-5bc6-477c-bf21-6842e1b3a9db nodeName:}" failed. No retries permitted until 2025-11-27 10:35:02.414429738 +0000 UTC m=+807.001020502 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert") pod "infra-operator-controller-manager-57548d458d-55nb9" (UID: "60db26f8-5bc6-477c-bf21-6842e1b3a9db") : secret "infra-operator-webhook-server-cert" not found Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.920589 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.926177 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.931656 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-68b9b" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.940040 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m"] Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.944895 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.952772 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-cvc6d" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.958834 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.973650 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88wlx\" (UniqueName: \"kubernetes.io/projected/60db26f8-5bc6-477c-bf21-6842e1b3a9db-kube-api-access-88wlx\") pod \"infra-operator-controller-manager-57548d458d-55nb9\" (UID: \"60db26f8-5bc6-477c-bf21-6842e1b3a9db\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:01 crc kubenswrapper[4806]: I1127 10:35:01.973659 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:01.999660 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.015212 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbhpg\" (UniqueName: \"kubernetes.io/projected/856215a1-cdc4-49fb-a522-0c2ca14cd757-kube-api-access-wbhpg\") pod \"keystone-operator-controller-manager-7b4567c7cf-m6mcb\" (UID: \"856215a1-cdc4-49fb-a522-0c2ca14cd757\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.015305 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpwsf\" (UniqueName: \"kubernetes.io/projected/4e3d3cf9-5e6d-4963-a2b2-731f4fb5ecb2-kube-api-access-vpwsf\") pod \"manila-operator-controller-manager-5d499bf58b-4vz5m\" (UID: \"4e3d3cf9-5e6d-4963-a2b2-731f4fb5ecb2\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.015424 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2xk5\" (UniqueName: \"kubernetes.io/projected/609248ce-9b50-48c5-90bd-aa3d2c8e2401-kube-api-access-q2xk5\") pod \"ironic-operator-controller-manager-67cb4dc6d4-nkrwj\" (UID: \"609248ce-9b50-48c5-90bd-aa3d2c8e2401\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.016323 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.059142 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.060063 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.060254 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.069404 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.070455 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.080313 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.080857 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-glwmg" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.088484 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-rpc8c" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.098925 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.100162 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.126043 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-wdspm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.127979 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2xk5\" (UniqueName: \"kubernetes.io/projected/609248ce-9b50-48c5-90bd-aa3d2c8e2401-kube-api-access-q2xk5\") pod \"ironic-operator-controller-manager-67cb4dc6d4-nkrwj\" (UID: \"609248ce-9b50-48c5-90bd-aa3d2c8e2401\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.128025 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbhpg\" (UniqueName: \"kubernetes.io/projected/856215a1-cdc4-49fb-a522-0c2ca14cd757-kube-api-access-wbhpg\") pod \"keystone-operator-controller-manager-7b4567c7cf-m6mcb\" (UID: \"856215a1-cdc4-49fb-a522-0c2ca14cd757\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.128062 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpwsf\" (UniqueName: \"kubernetes.io/projected/4e3d3cf9-5e6d-4963-a2b2-731f4fb5ecb2-kube-api-access-vpwsf\") pod \"manila-operator-controller-manager-5d499bf58b-4vz5m\" (UID: \"4e3d3cf9-5e6d-4963-a2b2-731f4fb5ecb2\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.128111 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g49ph\" (UniqueName: \"kubernetes.io/projected/4740d40d-863a-40a0-a46f-b515f10e4733-kube-api-access-g49ph\") pod \"neutron-operator-controller-manager-6fdcddb789-tdgzm\" (UID: \"4740d40d-863a-40a0-a46f-b515f10e4733\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.128136 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxpd8\" (UniqueName: \"kubernetes.io/projected/671409ea-bf88-4550-a3f7-1fb63a3e89ed-kube-api-access-zxpd8\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-j2x9t\" (UID: \"671409ea-bf88-4550-a3f7-1fb63a3e89ed\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.139566 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.159919 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.191893 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2xk5\" (UniqueName: \"kubernetes.io/projected/609248ce-9b50-48c5-90bd-aa3d2c8e2401-kube-api-access-q2xk5\") pod \"ironic-operator-controller-manager-67cb4dc6d4-nkrwj\" (UID: \"609248ce-9b50-48c5-90bd-aa3d2c8e2401\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.191997 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbhpg\" (UniqueName: \"kubernetes.io/projected/856215a1-cdc4-49fb-a522-0c2ca14cd757-kube-api-access-wbhpg\") pod \"keystone-operator-controller-manager-7b4567c7cf-m6mcb\" (UID: \"856215a1-cdc4-49fb-a522-0c2ca14cd757\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.202179 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.203381 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpwsf\" (UniqueName: \"kubernetes.io/projected/4e3d3cf9-5e6d-4963-a2b2-731f4fb5ecb2-kube-api-access-vpwsf\") pod \"manila-operator-controller-manager-5d499bf58b-4vz5m\" (UID: \"4e3d3cf9-5e6d-4963-a2b2-731f4fb5ecb2\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.209068 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.214818 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.226268 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.229820 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztb4v\" (UniqueName: \"kubernetes.io/projected/35e9f707-4424-43fc-8d97-bb804524a180-kube-api-access-ztb4v\") pod \"nova-operator-controller-manager-79556f57fc-jtsbv\" (UID: \"35e9f707-4424-43fc-8d97-bb804524a180\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.232637 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g49ph\" (UniqueName: \"kubernetes.io/projected/4740d40d-863a-40a0-a46f-b515f10e4733-kube-api-access-g49ph\") pod \"neutron-operator-controller-manager-6fdcddb789-tdgzm\" (UID: \"4740d40d-863a-40a0-a46f-b515f10e4733\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.233141 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxpd8\" (UniqueName: \"kubernetes.io/projected/671409ea-bf88-4550-a3f7-1fb63a3e89ed-kube-api-access-zxpd8\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-j2x9t\" (UID: \"671409ea-bf88-4550-a3f7-1fb63a3e89ed\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.248054 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.257701 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-5vm5p" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.271864 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g49ph\" (UniqueName: \"kubernetes.io/projected/4740d40d-863a-40a0-a46f-b515f10e4733-kube-api-access-g49ph\") pod \"neutron-operator-controller-manager-6fdcddb789-tdgzm\" (UID: \"4740d40d-863a-40a0-a46f-b515f10e4733\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.284721 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.290703 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.294497 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-skg76" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.294664 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.302449 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.303999 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.307138 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-jxm4b" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.307778 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.309168 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxpd8\" (UniqueName: \"kubernetes.io/projected/671409ea-bf88-4550-a3f7-1fb63a3e89ed-kube-api-access-zxpd8\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-j2x9t\" (UID: \"671409ea-bf88-4550-a3f7-1fb63a3e89ed\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.336322 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.336721 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztb4v\" (UniqueName: \"kubernetes.io/projected/35e9f707-4424-43fc-8d97-bb804524a180-kube-api-access-ztb4v\") pod \"nova-operator-controller-manager-79556f57fc-jtsbv\" (UID: \"35e9f707-4424-43fc-8d97-bb804524a180\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.336807 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmhf5\" (UniqueName: \"kubernetes.io/projected/e5d637d7-a046-4026-a157-abcb8d99ba3b-kube-api-access-cmhf5\") pod \"octavia-operator-controller-manager-64cdc6ff96-hpmjm\" (UID: \"e5d637d7-a046-4026-a157-abcb8d99ba3b\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.351334 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.361980 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.371112 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztb4v\" (UniqueName: \"kubernetes.io/projected/35e9f707-4424-43fc-8d97-bb804524a180-kube-api-access-ztb4v\") pod \"nova-operator-controller-manager-79556f57fc-jtsbv\" (UID: \"35e9f707-4424-43fc-8d97-bb804524a180\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.389275 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.395147 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.402695 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-5cvrf" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.425626 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.426761 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.431677 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-ttzpx" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.432036 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.440320 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert\") pod \"infra-operator-controller-manager-57548d458d-55nb9\" (UID: \"60db26f8-5bc6-477c-bf21-6842e1b3a9db\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.440365 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmhf5\" (UniqueName: \"kubernetes.io/projected/e5d637d7-a046-4026-a157-abcb8d99ba3b-kube-api-access-cmhf5\") pod \"octavia-operator-controller-manager-64cdc6ff96-hpmjm\" (UID: \"e5d637d7-a046-4026-a157-abcb8d99ba3b\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.440396 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh\" (UID: \"55fdd641-f8c1-417f-8abb-8adfc95d7008\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.440426 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp46x\" (UniqueName: \"kubernetes.io/projected/55fdd641-f8c1-417f-8abb-8adfc95d7008-kube-api-access-dp46x\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh\" (UID: \"55fdd641-f8c1-417f-8abb-8adfc95d7008\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.440446 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc9vt\" (UniqueName: \"kubernetes.io/projected/159a9bdd-85be-42dc-8d60-19c3eb1a90e4-kube-api-access-wc9vt\") pod \"ovn-operator-controller-manager-56897c768d-82qfv\" (UID: \"159a9bdd-85be-42dc-8d60-19c3eb1a90e4\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv" Nov 27 10:35:02 crc kubenswrapper[4806]: E1127 10:35:02.440585 4806 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 10:35:02 crc kubenswrapper[4806]: E1127 10:35:02.440625 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert podName:60db26f8-5bc6-477c-bf21-6842e1b3a9db nodeName:}" failed. No retries permitted until 2025-11-27 10:35:03.440611402 +0000 UTC m=+808.027202156 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert") pod "infra-operator-controller-manager-57548d458d-55nb9" (UID: "60db26f8-5bc6-477c-bf21-6842e1b3a9db") : secret "infra-operator-webhook-server-cert" not found Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.444916 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.463476 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.489061 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.490103 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmhf5\" (UniqueName: \"kubernetes.io/projected/e5d637d7-a046-4026-a157-abcb8d99ba3b-kube-api-access-cmhf5\") pod \"octavia-operator-controller-manager-64cdc6ff96-hpmjm\" (UID: \"e5d637d7-a046-4026-a157-abcb8d99ba3b\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.558819 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp46x\" (UniqueName: \"kubernetes.io/projected/55fdd641-f8c1-417f-8abb-8adfc95d7008-kube-api-access-dp46x\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh\" (UID: \"55fdd641-f8c1-417f-8abb-8adfc95d7008\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.558885 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc9vt\" (UniqueName: \"kubernetes.io/projected/159a9bdd-85be-42dc-8d60-19c3eb1a90e4-kube-api-access-wc9vt\") pod \"ovn-operator-controller-manager-56897c768d-82qfv\" (UID: \"159a9bdd-85be-42dc-8d60-19c3eb1a90e4\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.559223 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh\" (UID: \"55fdd641-f8c1-417f-8abb-8adfc95d7008\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:02 crc kubenswrapper[4806]: E1127 10:35:02.559956 4806 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 10:35:02 crc kubenswrapper[4806]: E1127 10:35:02.560018 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert podName:55fdd641-f8c1-417f-8abb-8adfc95d7008 nodeName:}" failed. No retries permitted until 2025-11-27 10:35:03.060001581 +0000 UTC m=+807.646592345 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" (UID: "55fdd641-f8c1-417f-8abb-8adfc95d7008") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.569968 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.570624 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.595526 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.624772 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc9vt\" (UniqueName: \"kubernetes.io/projected/159a9bdd-85be-42dc-8d60-19c3eb1a90e4-kube-api-access-wc9vt\") pod \"ovn-operator-controller-manager-56897c768d-82qfv\" (UID: \"159a9bdd-85be-42dc-8d60-19c3eb1a90e4\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.625886 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp46x\" (UniqueName: \"kubernetes.io/projected/55fdd641-f8c1-417f-8abb-8adfc95d7008-kube-api-access-dp46x\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh\" (UID: \"55fdd641-f8c1-417f-8abb-8adfc95d7008\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.641201 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.653321 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-qpvwv" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.673044 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bm7m\" (UniqueName: \"kubernetes.io/projected/92d56580-f256-46f7-81fa-e370048b4d0f-kube-api-access-4bm7m\") pod \"placement-operator-controller-manager-57988cc5b5-hrbw7\" (UID: \"92d56580-f256-46f7-81fa-e370048b4d0f\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.673153 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7x8v\" (UniqueName: \"kubernetes.io/projected/f77adb32-dfac-48de-b4a8-c9263cf5729d-kube-api-access-j7x8v\") pod \"swift-operator-controller-manager-d77b94747-xs5lm\" (UID: \"f77adb32-dfac-48de-b4a8-c9263cf5729d\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.681164 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.715761 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.788776 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7x8v\" (UniqueName: \"kubernetes.io/projected/f77adb32-dfac-48de-b4a8-c9263cf5729d-kube-api-access-j7x8v\") pod \"swift-operator-controller-manager-d77b94747-xs5lm\" (UID: \"f77adb32-dfac-48de-b4a8-c9263cf5729d\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.788923 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bm7m\" (UniqueName: \"kubernetes.io/projected/92d56580-f256-46f7-81fa-e370048b4d0f-kube-api-access-4bm7m\") pod \"placement-operator-controller-manager-57988cc5b5-hrbw7\" (UID: \"92d56580-f256-46f7-81fa-e370048b4d0f\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.788980 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n6c5\" (UniqueName: \"kubernetes.io/projected/d1ceaced-2100-4952-9bc8-cbd5c82886c5-kube-api-access-7n6c5\") pod \"telemetry-operator-controller-manager-76cc84c6bb-fjhns\" (UID: \"d1ceaced-2100-4952-9bc8-cbd5c82886c5\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.832154 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.833835 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.833920 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.836889 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-6wlwk" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.858292 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.859570 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.862868 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-z7ftl" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.888823 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bm7m\" (UniqueName: \"kubernetes.io/projected/92d56580-f256-46f7-81fa-e370048b4d0f-kube-api-access-4bm7m\") pod \"placement-operator-controller-manager-57988cc5b5-hrbw7\" (UID: \"92d56580-f256-46f7-81fa-e370048b4d0f\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.889052 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.891110 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg9mh\" (UniqueName: \"kubernetes.io/projected/08fa2215-67c9-4110-b54d-7a18e1641d88-kube-api-access-pg9mh\") pod \"test-operator-controller-manager-5cd6c7f4c8-6fkj4\" (UID: \"08fa2215-67c9-4110-b54d-7a18e1641d88\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.891167 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n6c5\" (UniqueName: \"kubernetes.io/projected/d1ceaced-2100-4952-9bc8-cbd5c82886c5-kube-api-access-7n6c5\") pod \"telemetry-operator-controller-manager-76cc84c6bb-fjhns\" (UID: \"d1ceaced-2100-4952-9bc8-cbd5c82886c5\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.891273 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv6lv\" (UniqueName: \"kubernetes.io/projected/b02aea7d-ec4b-425f-8d13-6205c91ee589-kube-api-access-kv6lv\") pod \"watcher-operator-controller-manager-656dcb59d4-v7tst\" (UID: \"b02aea7d-ec4b-425f-8d13-6205c91ee589\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.904190 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7x8v\" (UniqueName: \"kubernetes.io/projected/f77adb32-dfac-48de-b4a8-c9263cf5729d-kube-api-access-j7x8v\") pod \"swift-operator-controller-manager-d77b94747-xs5lm\" (UID: \"f77adb32-dfac-48de-b4a8-c9263cf5729d\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.923541 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n6c5\" (UniqueName: \"kubernetes.io/projected/d1ceaced-2100-4952-9bc8-cbd5c82886c5-kube-api-access-7n6c5\") pod \"telemetry-operator-controller-manager-76cc84c6bb-fjhns\" (UID: \"d1ceaced-2100-4952-9bc8-cbd5c82886c5\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.945478 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.947397 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.951223 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.954851 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.955182 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-4h7nk" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.955378 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.972459 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bz87k"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.973676 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bz87k" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.978645 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-ksk7v" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.983643 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bz87k"] Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.992732 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv6lv\" (UniqueName: \"kubernetes.io/projected/b02aea7d-ec4b-425f-8d13-6205c91ee589-kube-api-access-kv6lv\") pod \"watcher-operator-controller-manager-656dcb59d4-v7tst\" (UID: \"b02aea7d-ec4b-425f-8d13-6205c91ee589\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst" Nov 27 10:35:02 crc kubenswrapper[4806]: I1127 10:35:02.992804 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg9mh\" (UniqueName: \"kubernetes.io/projected/08fa2215-67c9-4110-b54d-7a18e1641d88-kube-api-access-pg9mh\") pod \"test-operator-controller-manager-5cd6c7f4c8-6fkj4\" (UID: \"08fa2215-67c9-4110-b54d-7a18e1641d88\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.038920 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv6lv\" (UniqueName: \"kubernetes.io/projected/b02aea7d-ec4b-425f-8d13-6205c91ee589-kube-api-access-kv6lv\") pod \"watcher-operator-controller-manager-656dcb59d4-v7tst\" (UID: \"b02aea7d-ec4b-425f-8d13-6205c91ee589\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.043822 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg9mh\" (UniqueName: \"kubernetes.io/projected/08fa2215-67c9-4110-b54d-7a18e1641d88-kube-api-access-pg9mh\") pod \"test-operator-controller-manager-5cd6c7f4c8-6fkj4\" (UID: \"08fa2215-67c9-4110-b54d-7a18e1641d88\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.056586 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.077540 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.093977 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh\" (UID: \"55fdd641-f8c1-417f-8abb-8adfc95d7008\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.094054 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt8st\" (UniqueName: \"kubernetes.io/projected/3104471d-6a15-4454-863e-b5754f748043-kube-api-access-lt8st\") pod \"rabbitmq-cluster-operator-manager-668c99d594-bz87k\" (UID: \"3104471d-6a15-4454-863e-b5754f748043\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bz87k" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.094078 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.094125 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.094147 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6mxs\" (UniqueName: \"kubernetes.io/projected/a11f64de-02ce-4b43-9ed8-54d0c844e48e-kube-api-access-f6mxs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:03 crc kubenswrapper[4806]: E1127 10:35:03.094317 4806 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 10:35:03 crc kubenswrapper[4806]: E1127 10:35:03.094360 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert podName:55fdd641-f8c1-417f-8abb-8adfc95d7008 nodeName:}" failed. No retries permitted until 2025-11-27 10:35:04.094345122 +0000 UTC m=+808.680935886 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" (UID: "55fdd641-f8c1-417f-8abb-8adfc95d7008") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.132952 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.150204 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.195380 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.195428 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6mxs\" (UniqueName: \"kubernetes.io/projected/a11f64de-02ce-4b43-9ed8-54d0c844e48e-kube-api-access-f6mxs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.195538 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt8st\" (UniqueName: \"kubernetes.io/projected/3104471d-6a15-4454-863e-b5754f748043-kube-api-access-lt8st\") pod \"rabbitmq-cluster-operator-manager-668c99d594-bz87k\" (UID: \"3104471d-6a15-4454-863e-b5754f748043\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bz87k" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.195559 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:03 crc kubenswrapper[4806]: E1127 10:35:03.195698 4806 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 10:35:03 crc kubenswrapper[4806]: E1127 10:35:03.195748 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs podName:a11f64de-02ce-4b43-9ed8-54d0c844e48e nodeName:}" failed. No retries permitted until 2025-11-27 10:35:03.695733159 +0000 UTC m=+808.282323923 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs") pod "openstack-operator-controller-manager-7fbdc78c67-7th5m" (UID: "a11f64de-02ce-4b43-9ed8-54d0c844e48e") : secret "webhook-server-cert" not found Nov 27 10:35:03 crc kubenswrapper[4806]: E1127 10:35:03.196037 4806 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 10:35:03 crc kubenswrapper[4806]: E1127 10:35:03.196067 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs podName:a11f64de-02ce-4b43-9ed8-54d0c844e48e nodeName:}" failed. No retries permitted until 2025-11-27 10:35:03.696058329 +0000 UTC m=+808.282649093 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs") pod "openstack-operator-controller-manager-7fbdc78c67-7th5m" (UID: "a11f64de-02ce-4b43-9ed8-54d0c844e48e") : secret "metrics-server-cert" not found Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.218799 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6mxs\" (UniqueName: \"kubernetes.io/projected/a11f64de-02ce-4b43-9ed8-54d0c844e48e-kube-api-access-f6mxs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.225126 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt8st\" (UniqueName: \"kubernetes.io/projected/3104471d-6a15-4454-863e-b5754f748043-kube-api-access-lt8st\") pod \"rabbitmq-cluster-operator-manager-668c99d594-bz87k\" (UID: \"3104471d-6a15-4454-863e-b5754f748043\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bz87k" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.266672 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26"] Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.271434 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2"] Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.284821 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.294768 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-ct9ps"] Nov 27 10:35:03 crc kubenswrapper[4806]: W1127 10:35:03.297437 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96cf5347_61b7_4ecb_8093_59b25482cd54.slice/crio-ec9ed5888eb0225f777f1f008ec901edca20a5dfebdce47cfb4704d994b399c0 WatchSource:0}: Error finding container ec9ed5888eb0225f777f1f008ec901edca20a5dfebdce47cfb4704d994b399c0: Status 404 returned error can't find the container with id ec9ed5888eb0225f777f1f008ec901edca20a5dfebdce47cfb4704d994b399c0 Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.314750 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8"] Nov 27 10:35:03 crc kubenswrapper[4806]: W1127 10:35:03.383906 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93cca2e7_6982_42d3_8848_10e4f7a70509.slice/crio-6f22411ed4d8128ed28d8ddb9ab47b1d09dd441c10f2f7e6355bf387a0775dee WatchSource:0}: Error finding container 6f22411ed4d8128ed28d8ddb9ab47b1d09dd441c10f2f7e6355bf387a0775dee: Status 404 returned error can't find the container with id 6f22411ed4d8128ed28d8ddb9ab47b1d09dd441c10f2f7e6355bf387a0775dee Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.386199 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67"] Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.386405 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bz87k" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.412560 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6"] Nov 27 10:35:03 crc kubenswrapper[4806]: W1127 10:35:03.491268 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6163410_3453_43c3_8cd7_5a354e447025.slice/crio-c8da7f496bdc20af4e96f7206f69fd0b30b02369f5a7d4ba71a330a0a9e8d1d1 WatchSource:0}: Error finding container c8da7f496bdc20af4e96f7206f69fd0b30b02369f5a7d4ba71a330a0a9e8d1d1: Status 404 returned error can't find the container with id c8da7f496bdc20af4e96f7206f69fd0b30b02369f5a7d4ba71a330a0a9e8d1d1 Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.501425 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert\") pod \"infra-operator-controller-manager-57548d458d-55nb9\" (UID: \"60db26f8-5bc6-477c-bf21-6842e1b3a9db\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:03 crc kubenswrapper[4806]: E1127 10:35:03.501637 4806 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 10:35:03 crc kubenswrapper[4806]: E1127 10:35:03.501687 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert podName:60db26f8-5bc6-477c-bf21-6842e1b3a9db nodeName:}" failed. No retries permitted until 2025-11-27 10:35:05.501672982 +0000 UTC m=+810.088263746 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert") pod "infra-operator-controller-manager-57548d458d-55nb9" (UID: "60db26f8-5bc6-477c-bf21-6842e1b3a9db") : secret "infra-operator-webhook-server-cert" not found Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.695184 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8" event={"ID":"94f64657-d78c-4f52-8aa1-b23b61d5af68","Type":"ContainerStarted","Data":"22944f902899c513f8f3dfe4ae232558efabd561eff60faeb7fec695fd85e0f2"} Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.698764 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-ct9ps" event={"ID":"93cca2e7-6982-42d3-8848-10e4f7a70509","Type":"ContainerStarted","Data":"6f22411ed4d8128ed28d8ddb9ab47b1d09dd441c10f2f7e6355bf387a0775dee"} Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.706671 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.706736 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:03 crc kubenswrapper[4806]: E1127 10:35:03.706853 4806 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 10:35:03 crc kubenswrapper[4806]: E1127 10:35:03.706900 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs podName:a11f64de-02ce-4b43-9ed8-54d0c844e48e nodeName:}" failed. No retries permitted until 2025-11-27 10:35:04.706885954 +0000 UTC m=+809.293476718 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs") pod "openstack-operator-controller-manager-7fbdc78c67-7th5m" (UID: "a11f64de-02ce-4b43-9ed8-54d0c844e48e") : secret "metrics-server-cert" not found Nov 27 10:35:03 crc kubenswrapper[4806]: E1127 10:35:03.707194 4806 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 10:35:03 crc kubenswrapper[4806]: E1127 10:35:03.707269 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs podName:a11f64de-02ce-4b43-9ed8-54d0c844e48e nodeName:}" failed. No retries permitted until 2025-11-27 10:35:04.707225384 +0000 UTC m=+809.293816148 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs") pod "openstack-operator-controller-manager-7fbdc78c67-7th5m" (UID: "a11f64de-02ce-4b43-9ed8-54d0c844e48e") : secret "webhook-server-cert" not found Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.707388 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2" event={"ID":"96cf5347-61b7-4ecb-8093-59b25482cd54","Type":"ContainerStarted","Data":"ec9ed5888eb0225f777f1f008ec901edca20a5dfebdce47cfb4704d994b399c0"} Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.736766 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67" event={"ID":"b6163410-3453-43c3-8cd7-5a354e447025","Type":"ContainerStarted","Data":"c8da7f496bdc20af4e96f7206f69fd0b30b02369f5a7d4ba71a330a0a9e8d1d1"} Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.745558 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6" event={"ID":"33c0104b-4a8a-46fa-bd83-b42f86f96789","Type":"ContainerStarted","Data":"0d57ddd2753fc471b3a7db7a321a0d52cd88a533f6904af0c17941445ce2c217"} Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.758188 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26" event={"ID":"5b227bdd-2c69-4053-a116-8085cd157a93","Type":"ContainerStarted","Data":"9859e3459bff9e28a786a06da18ca2cd68a62581c8ed86bca8bcf32cd9503af4"} Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.760875 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m"] Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.808153 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm"] Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.838828 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj"] Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.972438 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv"] Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.983783 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb"] Nov 27 10:35:03 crc kubenswrapper[4806]: I1127 10:35:03.997550 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst"] Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.011109 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t"] Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.031290 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm"] Nov 27 10:35:04 crc kubenswrapper[4806]: W1127 10:35:04.047345 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5d637d7_a046_4026_a157_abcb8d99ba3b.slice/crio-6ede4b45245935bcfe1f5cc11244a84f3bcbff731e8a5755cbdcc26f1d3381d0 WatchSource:0}: Error finding container 6ede4b45245935bcfe1f5cc11244a84f3bcbff731e8a5755cbdcc26f1d3381d0: Status 404 returned error can't find the container with id 6ede4b45245935bcfe1f5cc11244a84f3bcbff731e8a5755cbdcc26f1d3381d0 Nov 27 10:35:04 crc kubenswrapper[4806]: W1127 10:35:04.047696 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35e9f707_4424_43fc_8d97_bb804524a180.slice/crio-7f37931f2961b2e1337b81484d6a8ae68578caed2379fbc7584824b1e2cc1a6d WatchSource:0}: Error finding container 7f37931f2961b2e1337b81484d6a8ae68578caed2379fbc7584824b1e2cc1a6d: Status 404 returned error can't find the container with id 7f37931f2961b2e1337b81484d6a8ae68578caed2379fbc7584824b1e2cc1a6d Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.050115 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:ddc8a82f05930db8ee7a8d6d189b5a66373060656e4baf71ac302f89c477da4c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cmhf5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-64cdc6ff96-hpmjm_openstack-operators(e5d637d7-a046-4026-a157-abcb8d99ba3b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.050488 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ztb4v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-jtsbv_openstack-operators(35e9f707-4424-43fc-8d97-bb804524a180): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.057777 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cmhf5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-64cdc6ff96-hpmjm_openstack-operators(e5d637d7-a046-4026-a157-abcb8d99ba3b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.058126 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ztb4v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-jtsbv_openstack-operators(35e9f707-4424-43fc-8d97-bb804524a180): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.059327 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" podUID="e5d637d7-a046-4026-a157-abcb8d99ba3b" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.059388 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" podUID="35e9f707-4424-43fc-8d97-bb804524a180" Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.061851 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv"] Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.114093 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh\" (UID: \"55fdd641-f8c1-417f-8abb-8adfc95d7008\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.114369 4806 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.114423 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert podName:55fdd641-f8c1-417f-8abb-8adfc95d7008 nodeName:}" failed. No retries permitted until 2025-11-27 10:35:06.114406498 +0000 UTC m=+810.700997262 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" (UID: "55fdd641-f8c1-417f-8abb-8adfc95d7008") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.158872 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4"] Nov 27 10:35:04 crc kubenswrapper[4806]: W1127 10:35:04.182207 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08fa2215_67c9_4110_b54d_7a18e1641d88.slice/crio-4492503f0a44165f04bb9454a45611e9f4e794a5ce03fcb63557f16fea1e2e74 WatchSource:0}: Error finding container 4492503f0a44165f04bb9454a45611e9f4e794a5ce03fcb63557f16fea1e2e74: Status 404 returned error can't find the container with id 4492503f0a44165f04bb9454a45611e9f4e794a5ce03fcb63557f16fea1e2e74 Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.185466 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm"] Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.208087 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7"] Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.214217 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns"] Nov 27 10:35:04 crc kubenswrapper[4806]: W1127 10:35:04.220067 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf77adb32_dfac_48de_b4a8_c9263cf5729d.slice/crio-9b06803851e5991e3bacd5a1b5c454780f3a411f70a428b5cd10b06515f74710 WatchSource:0}: Error finding container 9b06803851e5991e3bacd5a1b5c454780f3a411f70a428b5cd10b06515f74710: Status 404 returned error can't find the container with id 9b06803851e5991e3bacd5a1b5c454780f3a411f70a428b5cd10b06515f74710 Nov 27 10:35:04 crc kubenswrapper[4806]: W1127 10:35:04.220876 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92d56580_f256_46f7_81fa_e370048b4d0f.slice/crio-f6a97fa8c29b5ae25cb46ab9581fe8e4d008556dcd6279e386a91b90100fc143 WatchSource:0}: Error finding container f6a97fa8c29b5ae25cb46ab9581fe8e4d008556dcd6279e386a91b90100fc143: Status 404 returned error can't find the container with id f6a97fa8c29b5ae25cb46ab9581fe8e4d008556dcd6279e386a91b90100fc143 Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.232603 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4bm7m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-57988cc5b5-hrbw7_openstack-operators(92d56580-f256-46f7-81fa-e370048b4d0f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.233509 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j7x8v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d77b94747-xs5lm_openstack-operators(f77adb32-dfac-48de-b4a8-c9263cf5729d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.236038 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4bm7m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-57988cc5b5-hrbw7_openstack-operators(92d56580-f256-46f7-81fa-e370048b4d0f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.236864 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j7x8v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d77b94747-xs5lm_openstack-operators(f77adb32-dfac-48de-b4a8-c9263cf5729d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.237596 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" podUID="92d56580-f256-46f7-81fa-e370048b4d0f" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.238865 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" podUID="f77adb32-dfac-48de-b4a8-c9263cf5729d" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.239353 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7n6c5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-fjhns_openstack-operators(d1ceaced-2100-4952-9bc8-cbd5c82886c5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.244800 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7n6c5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-fjhns_openstack-operators(d1ceaced-2100-4952-9bc8-cbd5c82886c5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.245961 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" podUID="d1ceaced-2100-4952-9bc8-cbd5c82886c5" Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.327210 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bz87k"] Nov 27 10:35:04 crc kubenswrapper[4806]: W1127 10:35:04.333279 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3104471d_6a15_4454_863e_b5754f748043.slice/crio-d3905b5cfa7e378fef3b0e29bcb48a56d9fe1fe01c0726186cc20d7607e56ce2 WatchSource:0}: Error finding container d3905b5cfa7e378fef3b0e29bcb48a56d9fe1fe01c0726186cc20d7607e56ce2: Status 404 returned error can't find the container with id d3905b5cfa7e378fef3b0e29bcb48a56d9fe1fe01c0726186cc20d7607e56ce2 Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.730946 4806 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.731094 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs podName:a11f64de-02ce-4b43-9ed8-54d0c844e48e nodeName:}" failed. No retries permitted until 2025-11-27 10:35:06.731069485 +0000 UTC m=+811.317660249 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs") pod "openstack-operator-controller-manager-7fbdc78c67-7th5m" (UID: "a11f64de-02ce-4b43-9ed8-54d0c844e48e") : secret "webhook-server-cert" not found Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.730774 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.731711 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.731805 4806 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.731852 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs podName:a11f64de-02ce-4b43-9ed8-54d0c844e48e nodeName:}" failed. No retries permitted until 2025-11-27 10:35:06.731843627 +0000 UTC m=+811.318434391 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs") pod "openstack-operator-controller-manager-7fbdc78c67-7th5m" (UID: "a11f64de-02ce-4b43-9ed8-54d0c844e48e") : secret "metrics-server-cert" not found Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.770810 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" event={"ID":"35e9f707-4424-43fc-8d97-bb804524a180","Type":"ContainerStarted","Data":"7f37931f2961b2e1337b81484d6a8ae68578caed2379fbc7584824b1e2cc1a6d"} Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.781603 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" podUID="35e9f707-4424-43fc-8d97-bb804524a180" Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.782684 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bz87k" event={"ID":"3104471d-6a15-4454-863e-b5754f748043","Type":"ContainerStarted","Data":"d3905b5cfa7e378fef3b0e29bcb48a56d9fe1fe01c0726186cc20d7607e56ce2"} Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.788706 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv" event={"ID":"159a9bdd-85be-42dc-8d60-19c3eb1a90e4","Type":"ContainerStarted","Data":"25f8b0f84b6229f214a875672483ba2e14502f5fac6865b35aab98666cc31e94"} Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.802580 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb" event={"ID":"856215a1-cdc4-49fb-a522-0c2ca14cd757","Type":"ContainerStarted","Data":"d54e207acba6a3fa9b64428aec5adb103ecf567befff52c00d79d36f9ee3ebfb"} Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.804558 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj" event={"ID":"609248ce-9b50-48c5-90bd-aa3d2c8e2401","Type":"ContainerStarted","Data":"b625a9c78f1b034ff3621f782ce4dc5ba3f96498365a39b34cddce9556597301"} Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.816989 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst" event={"ID":"b02aea7d-ec4b-425f-8d13-6205c91ee589","Type":"ContainerStarted","Data":"df269a28a47b23ef212a7a9f8e4baa642d231336bd2861c32afc8db9ff469011"} Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.820255 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" event={"ID":"92d56580-f256-46f7-81fa-e370048b4d0f","Type":"ContainerStarted","Data":"f6a97fa8c29b5ae25cb46ab9581fe8e4d008556dcd6279e386a91b90100fc143"} Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.825387 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" podUID="92d56580-f256-46f7-81fa-e370048b4d0f" Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.831837 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" event={"ID":"f77adb32-dfac-48de-b4a8-c9263cf5729d","Type":"ContainerStarted","Data":"9b06803851e5991e3bacd5a1b5c454780f3a411f70a428b5cd10b06515f74710"} Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.843376 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" event={"ID":"e5d637d7-a046-4026-a157-abcb8d99ba3b","Type":"ContainerStarted","Data":"6ede4b45245935bcfe1f5cc11244a84f3bcbff731e8a5755cbdcc26f1d3381d0"} Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.852005 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:ddc8a82f05930db8ee7a8d6d189b5a66373060656e4baf71ac302f89c477da4c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" podUID="e5d637d7-a046-4026-a157-abcb8d99ba3b" Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.863794 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" podUID="f77adb32-dfac-48de-b4a8-c9263cf5729d" Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.869109 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4" event={"ID":"08fa2215-67c9-4110-b54d-7a18e1641d88","Type":"ContainerStarted","Data":"4492503f0a44165f04bb9454a45611e9f4e794a5ce03fcb63557f16fea1e2e74"} Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.876695 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" event={"ID":"d1ceaced-2100-4952-9bc8-cbd5c82886c5","Type":"ContainerStarted","Data":"18669c0b5c20f41d1d46af946cd120e3f23db90639b4c8cff64013612d769e2b"} Nov 27 10:35:04 crc kubenswrapper[4806]: E1127 10:35:04.880742 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" podUID="d1ceaced-2100-4952-9bc8-cbd5c82886c5" Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.896481 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m" event={"ID":"4e3d3cf9-5e6d-4963-a2b2-731f4fb5ecb2","Type":"ContainerStarted","Data":"75d6798d58b74009b83bd50a8ebb2e51f98dbd5fb1f37512494c4fc0cf8f645f"} Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.899776 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t" event={"ID":"671409ea-bf88-4550-a3f7-1fb63a3e89ed","Type":"ContainerStarted","Data":"9d8f68ad70261e78b12e20e1d69a3cf1e7ecbbae950640013363434fc2537b56"} Nov 27 10:35:04 crc kubenswrapper[4806]: I1127 10:35:04.914349 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm" event={"ID":"4740d40d-863a-40a0-a46f-b515f10e4733","Type":"ContainerStarted","Data":"becb14df017b0fe4bd009a82f4888e1a90f97166d02527e08d8b25fb318fc7a1"} Nov 27 10:35:05 crc kubenswrapper[4806]: I1127 10:35:05.546671 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert\") pod \"infra-operator-controller-manager-57548d458d-55nb9\" (UID: \"60db26f8-5bc6-477c-bf21-6842e1b3a9db\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:05 crc kubenswrapper[4806]: E1127 10:35:05.547060 4806 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 10:35:05 crc kubenswrapper[4806]: E1127 10:35:05.547152 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert podName:60db26f8-5bc6-477c-bf21-6842e1b3a9db nodeName:}" failed. No retries permitted until 2025-11-27 10:35:09.547129893 +0000 UTC m=+814.133720657 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert") pod "infra-operator-controller-manager-57548d458d-55nb9" (UID: "60db26f8-5bc6-477c-bf21-6842e1b3a9db") : secret "infra-operator-webhook-server-cert" not found Nov 27 10:35:05 crc kubenswrapper[4806]: E1127 10:35:05.944624 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" podUID="35e9f707-4424-43fc-8d97-bb804524a180" Nov 27 10:35:05 crc kubenswrapper[4806]: E1127 10:35:05.944726 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" podUID="f77adb32-dfac-48de-b4a8-c9263cf5729d" Nov 27 10:35:05 crc kubenswrapper[4806]: E1127 10:35:05.945489 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" podUID="d1ceaced-2100-4952-9bc8-cbd5c82886c5" Nov 27 10:35:05 crc kubenswrapper[4806]: E1127 10:35:05.946604 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:ddc8a82f05930db8ee7a8d6d189b5a66373060656e4baf71ac302f89c477da4c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" podUID="e5d637d7-a046-4026-a157-abcb8d99ba3b" Nov 27 10:35:05 crc kubenswrapper[4806]: E1127 10:35:05.946649 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" podUID="92d56580-f256-46f7-81fa-e370048b4d0f" Nov 27 10:35:06 crc kubenswrapper[4806]: I1127 10:35:06.166134 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh\" (UID: \"55fdd641-f8c1-417f-8abb-8adfc95d7008\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:06 crc kubenswrapper[4806]: E1127 10:35:06.166634 4806 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 10:35:06 crc kubenswrapper[4806]: E1127 10:35:06.167886 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert podName:55fdd641-f8c1-417f-8abb-8adfc95d7008 nodeName:}" failed. No retries permitted until 2025-11-27 10:35:10.167855593 +0000 UTC m=+814.754446357 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" (UID: "55fdd641-f8c1-417f-8abb-8adfc95d7008") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 10:35:06 crc kubenswrapper[4806]: I1127 10:35:06.782387 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:06 crc kubenswrapper[4806]: I1127 10:35:06.782489 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:06 crc kubenswrapper[4806]: E1127 10:35:06.782733 4806 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 10:35:06 crc kubenswrapper[4806]: E1127 10:35:06.782810 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs podName:a11f64de-02ce-4b43-9ed8-54d0c844e48e nodeName:}" failed. No retries permitted until 2025-11-27 10:35:10.782784931 +0000 UTC m=+815.369375695 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs") pod "openstack-operator-controller-manager-7fbdc78c67-7th5m" (UID: "a11f64de-02ce-4b43-9ed8-54d0c844e48e") : secret "metrics-server-cert" not found Nov 27 10:35:06 crc kubenswrapper[4806]: E1127 10:35:06.782936 4806 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 10:35:06 crc kubenswrapper[4806]: E1127 10:35:06.782970 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs podName:a11f64de-02ce-4b43-9ed8-54d0c844e48e nodeName:}" failed. No retries permitted until 2025-11-27 10:35:10.782963016 +0000 UTC m=+815.369553770 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs") pod "openstack-operator-controller-manager-7fbdc78c67-7th5m" (UID: "a11f64de-02ce-4b43-9ed8-54d0c844e48e") : secret "webhook-server-cert" not found Nov 27 10:35:09 crc kubenswrapper[4806]: I1127 10:35:09.646887 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert\") pod \"infra-operator-controller-manager-57548d458d-55nb9\" (UID: \"60db26f8-5bc6-477c-bf21-6842e1b3a9db\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:09 crc kubenswrapper[4806]: E1127 10:35:09.647125 4806 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 27 10:35:09 crc kubenswrapper[4806]: E1127 10:35:09.647648 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert podName:60db26f8-5bc6-477c-bf21-6842e1b3a9db nodeName:}" failed. No retries permitted until 2025-11-27 10:35:17.647626373 +0000 UTC m=+822.234217127 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert") pod "infra-operator-controller-manager-57548d458d-55nb9" (UID: "60db26f8-5bc6-477c-bf21-6842e1b3a9db") : secret "infra-operator-webhook-server-cert" not found Nov 27 10:35:10 crc kubenswrapper[4806]: I1127 10:35:10.261384 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh\" (UID: \"55fdd641-f8c1-417f-8abb-8adfc95d7008\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:10 crc kubenswrapper[4806]: E1127 10:35:10.261784 4806 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 10:35:10 crc kubenswrapper[4806]: E1127 10:35:10.261842 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert podName:55fdd641-f8c1-417f-8abb-8adfc95d7008 nodeName:}" failed. No retries permitted until 2025-11-27 10:35:18.261824181 +0000 UTC m=+822.848414945 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" (UID: "55fdd641-f8c1-417f-8abb-8adfc95d7008") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 27 10:35:10 crc kubenswrapper[4806]: E1127 10:35:10.872984 4806 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 27 10:35:10 crc kubenswrapper[4806]: E1127 10:35:10.873093 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs podName:a11f64de-02ce-4b43-9ed8-54d0c844e48e nodeName:}" failed. No retries permitted until 2025-11-27 10:35:18.873068166 +0000 UTC m=+823.459658940 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs") pod "openstack-operator-controller-manager-7fbdc78c67-7th5m" (UID: "a11f64de-02ce-4b43-9ed8-54d0c844e48e") : secret "metrics-server-cert" not found Nov 27 10:35:10 crc kubenswrapper[4806]: I1127 10:35:10.872797 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:10 crc kubenswrapper[4806]: I1127 10:35:10.873814 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:10 crc kubenswrapper[4806]: E1127 10:35:10.873922 4806 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 27 10:35:10 crc kubenswrapper[4806]: E1127 10:35:10.873966 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs podName:a11f64de-02ce-4b43-9ed8-54d0c844e48e nodeName:}" failed. No retries permitted until 2025-11-27 10:35:18.873955911 +0000 UTC m=+823.460546695 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs") pod "openstack-operator-controller-manager-7fbdc78c67-7th5m" (UID: "a11f64de-02ce-4b43-9ed8-54d0c844e48e") : secret "webhook-server-cert" not found Nov 27 10:35:16 crc kubenswrapper[4806]: I1127 10:35:16.005141 4806 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 27 10:35:16 crc kubenswrapper[4806]: E1127 10:35:16.951967 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:45ae665ce2ea81aef212ee402cb02693ee49001a7c88c40c9598ff2859b838a2" Nov 27 10:35:16 crc kubenswrapper[4806]: E1127 10:35:16.952589 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:45ae665ce2ea81aef212ee402cb02693ee49001a7c88c40c9598ff2859b838a2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2m6zb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-589cbd6b5b-9jh67_openstack-operators(b6163410-3453-43c3-8cd7-5a354e447025): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:35:17 crc kubenswrapper[4806]: E1127 10:35:17.622661 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6" Nov 27 10:35:17 crc kubenswrapper[4806]: E1127 10:35:17.623278 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wc9vt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-56897c768d-82qfv_openstack-operators(159a9bdd-85be-42dc-8d60-19c3eb1a90e4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:35:17 crc kubenswrapper[4806]: I1127 10:35:17.707930 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert\") pod \"infra-operator-controller-manager-57548d458d-55nb9\" (UID: \"60db26f8-5bc6-477c-bf21-6842e1b3a9db\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:17 crc kubenswrapper[4806]: I1127 10:35:17.716420 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/60db26f8-5bc6-477c-bf21-6842e1b3a9db-cert\") pod \"infra-operator-controller-manager-57548d458d-55nb9\" (UID: \"60db26f8-5bc6-477c-bf21-6842e1b3a9db\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:17 crc kubenswrapper[4806]: I1127 10:35:17.745979 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:18 crc kubenswrapper[4806]: I1127 10:35:18.317161 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh\" (UID: \"55fdd641-f8c1-417f-8abb-8adfc95d7008\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:18 crc kubenswrapper[4806]: I1127 10:35:18.321933 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55fdd641-f8c1-417f-8abb-8adfc95d7008-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh\" (UID: \"55fdd641-f8c1-417f-8abb-8adfc95d7008\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:18 crc kubenswrapper[4806]: I1127 10:35:18.596606 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:18 crc kubenswrapper[4806]: I1127 10:35:18.926640 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:18 crc kubenswrapper[4806]: I1127 10:35:18.927326 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:18 crc kubenswrapper[4806]: I1127 10:35:18.934092 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-metrics-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:18 crc kubenswrapper[4806]: I1127 10:35:18.934237 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a11f64de-02ce-4b43-9ed8-54d0c844e48e-webhook-certs\") pod \"openstack-operator-controller-manager-7fbdc78c67-7th5m\" (UID: \"a11f64de-02ce-4b43-9ed8-54d0c844e48e\") " pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:18 crc kubenswrapper[4806]: I1127 10:35:18.968931 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:23 crc kubenswrapper[4806]: E1127 10:35:23.111297 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:2ee37ff474bee3203447df4f326a9279a515e770573153338296dd074722c677" Nov 27 10:35:23 crc kubenswrapper[4806]: E1127 10:35:23.111884 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:2ee37ff474bee3203447df4f326a9279a515e770573153338296dd074722c677,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7xngs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5b77f656f-czdn8_openstack-operators(94f64657-d78c-4f52-8aa1-b23b61d5af68): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:35:28 crc kubenswrapper[4806]: E1127 10:35:28.698426 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Nov 27 10:35:28 crc kubenswrapper[4806]: E1127 10:35:28.699264 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lt8st,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-bz87k_openstack-operators(3104471d-6a15-4454-863e-b5754f748043): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:35:28 crc kubenswrapper[4806]: E1127 10:35:28.700572 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bz87k" podUID="3104471d-6a15-4454-863e-b5754f748043" Nov 27 10:35:29 crc kubenswrapper[4806]: E1127 10:35:29.177333 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bz87k" podUID="3104471d-6a15-4454-863e-b5754f748043" Nov 27 10:35:30 crc kubenswrapper[4806]: E1127 10:35:30.026611 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711" Nov 27 10:35:30 crc kubenswrapper[4806]: E1127 10:35:30.026786 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:25faa5b0e4801d4d3b01a28b877ed3188eee71f33ad66f3c2e86b7921758e711,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wbhpg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7b4567c7cf-m6mcb_openstack-operators(856215a1-cdc4-49fb-a522-0c2ca14cd757): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:35:32 crc kubenswrapper[4806]: I1127 10:35:32.412105 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh"] Nov 27 10:35:32 crc kubenswrapper[4806]: I1127 10:35:32.485257 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-55nb9"] Nov 27 10:35:32 crc kubenswrapper[4806]: I1127 10:35:32.497851 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m"] Nov 27 10:35:33 crc kubenswrapper[4806]: I1127 10:35:33.212598 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m" event={"ID":"4e3d3cf9-5e6d-4963-a2b2-731f4fb5ecb2","Type":"ContainerStarted","Data":"864cc457edc75fdf5b64af98da17845fe22ad1113e6ab34f3f3b56ff2c4ddd2c"} Nov 27 10:35:33 crc kubenswrapper[4806]: I1127 10:35:33.215127 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst" event={"ID":"b02aea7d-ec4b-425f-8d13-6205c91ee589","Type":"ContainerStarted","Data":"9a250b94d35c150c48cdc21ccb46e5c839ff0bfc2981d682e09a3d2b616b5557"} Nov 27 10:35:36 crc kubenswrapper[4806]: W1127 10:35:36.047092 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55fdd641_f8c1_417f_8abb_8adfc95d7008.slice/crio-1b3120c72f0faa1493fdf61b2d2cd4e62faf132280cfe3482016c8a4addac0bd WatchSource:0}: Error finding container 1b3120c72f0faa1493fdf61b2d2cd4e62faf132280cfe3482016c8a4addac0bd: Status 404 returned error can't find the container with id 1b3120c72f0faa1493fdf61b2d2cd4e62faf132280cfe3482016c8a4addac0bd Nov 27 10:35:36 crc kubenswrapper[4806]: W1127 10:35:36.049060 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60db26f8_5bc6_477c_bf21_6842e1b3a9db.slice/crio-ea397901bbe4090a117a2078b2e479f11378d766730c00b6f5dc2a7be321b62b WatchSource:0}: Error finding container ea397901bbe4090a117a2078b2e479f11378d766730c00b6f5dc2a7be321b62b: Status 404 returned error can't find the container with id ea397901bbe4090a117a2078b2e479f11378d766730c00b6f5dc2a7be321b62b Nov 27 10:35:36 crc kubenswrapper[4806]: I1127 10:35:36.245754 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" event={"ID":"a11f64de-02ce-4b43-9ed8-54d0c844e48e","Type":"ContainerStarted","Data":"f9719f4829ff8ba25818785a5fa62e015550cb480929392ca5ad796e3b15285e"} Nov 27 10:35:36 crc kubenswrapper[4806]: I1127 10:35:36.249520 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" event={"ID":"f77adb32-dfac-48de-b4a8-c9263cf5729d","Type":"ContainerStarted","Data":"f67b8909727bc72dca62c381bb9a5657ca8d189182d175b71e119b2a5a6b9eef"} Nov 27 10:35:36 crc kubenswrapper[4806]: I1127 10:35:36.251715 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" event={"ID":"60db26f8-5bc6-477c-bf21-6842e1b3a9db","Type":"ContainerStarted","Data":"ea397901bbe4090a117a2078b2e479f11378d766730c00b6f5dc2a7be321b62b"} Nov 27 10:35:36 crc kubenswrapper[4806]: I1127 10:35:36.253957 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" event={"ID":"55fdd641-f8c1-417f-8abb-8adfc95d7008","Type":"ContainerStarted","Data":"1b3120c72f0faa1493fdf61b2d2cd4e62faf132280cfe3482016c8a4addac0bd"} Nov 27 10:35:37 crc kubenswrapper[4806]: I1127 10:35:37.291555 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm" event={"ID":"4740d40d-863a-40a0-a46f-b515f10e4733","Type":"ContainerStarted","Data":"e50a96672ca7ef4d98fcdc7ab7a57577fd9cb4205d7fa20c5490e3120fa51a3e"} Nov 27 10:35:37 crc kubenswrapper[4806]: I1127 10:35:37.313905 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26" event={"ID":"5b227bdd-2c69-4053-a116-8085cd157a93","Type":"ContainerStarted","Data":"cbd0c11970f9064413fe79cc5e98acfcc62664f3524d8dabfa4c1649f9a4c29b"} Nov 27 10:35:37 crc kubenswrapper[4806]: I1127 10:35:37.329527 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t" event={"ID":"671409ea-bf88-4550-a3f7-1fb63a3e89ed","Type":"ContainerStarted","Data":"9320195fe15eac9e61238d33c4e709bdf0c9596acbf32305f2f889c022ea56d4"} Nov 27 10:35:37 crc kubenswrapper[4806]: I1127 10:35:37.353970 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2" event={"ID":"96cf5347-61b7-4ecb-8093-59b25482cd54","Type":"ContainerStarted","Data":"d34c78395fd336ce65f211187ac9e9e80dceeccccb5b080cdaed7b4ae64cdf62"} Nov 27 10:35:37 crc kubenswrapper[4806]: I1127 10:35:37.366324 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6" event={"ID":"33c0104b-4a8a-46fa-bd83-b42f86f96789","Type":"ContainerStarted","Data":"fbcac566abc3eee3ea81a86e3e76750fdbba9a5920c602eddce33dab4a225a55"} Nov 27 10:35:37 crc kubenswrapper[4806]: I1127 10:35:37.398562 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj" event={"ID":"609248ce-9b50-48c5-90bd-aa3d2c8e2401","Type":"ContainerStarted","Data":"f5610c4b366ccb49aefb228567af66f62aec4997db83940c8344defb471e4075"} Nov 27 10:35:37 crc kubenswrapper[4806]: I1127 10:35:37.412416 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-ct9ps" event={"ID":"93cca2e7-6982-42d3-8848-10e4f7a70509","Type":"ContainerStarted","Data":"8ee1e79a3e8274546c900e1eaa94f75749fd2ca5c33a7ca274c023028cb65d36"} Nov 27 10:35:37 crc kubenswrapper[4806]: I1127 10:35:37.421137 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" event={"ID":"92d56580-f256-46f7-81fa-e370048b4d0f","Type":"ContainerStarted","Data":"a7667000db331e1894ebdd4c985862fbc72ba4772ade11a06c952c7668885aab"} Nov 27 10:35:38 crc kubenswrapper[4806]: I1127 10:35:38.436894 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" event={"ID":"d1ceaced-2100-4952-9bc8-cbd5c82886c5","Type":"ContainerStarted","Data":"7e8c5099487abf00e2e4767b978599c6f4445e26630ff210c2144baf0e993f39"} Nov 27 10:35:38 crc kubenswrapper[4806]: I1127 10:35:38.442888 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" event={"ID":"a11f64de-02ce-4b43-9ed8-54d0c844e48e","Type":"ContainerStarted","Data":"cccf246eba5b0141d4d5f441a7b1e3173743c6ed1af5b4658430689f143fe440"} Nov 27 10:35:38 crc kubenswrapper[4806]: I1127 10:35:38.443101 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:38 crc kubenswrapper[4806]: I1127 10:35:38.447431 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" event={"ID":"e5d637d7-a046-4026-a157-abcb8d99ba3b","Type":"ContainerStarted","Data":"f826d9a0591f7120fb985f6c737cf14b6d475378eb3f2cb590ebfe3cad6d2678"} Nov 27 10:35:38 crc kubenswrapper[4806]: I1127 10:35:38.455528 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4" event={"ID":"08fa2215-67c9-4110-b54d-7a18e1641d88","Type":"ContainerStarted","Data":"7a37adeb65dc6c69637758a09fd2c86f5f8883cfa5cf742786fef3c51f242306"} Nov 27 10:35:38 crc kubenswrapper[4806]: I1127 10:35:38.492341 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" podStartSLOduration=36.492323101 podStartE2EDuration="36.492323101s" podCreationTimestamp="2025-11-27 10:35:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:35:38.475530073 +0000 UTC m=+843.062120837" watchObservedRunningTime="2025-11-27 10:35:38.492323101 +0000 UTC m=+843.078913865" Nov 27 10:35:42 crc kubenswrapper[4806]: I1127 10:35:42.505745 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" event={"ID":"35e9f707-4424-43fc-8d97-bb804524a180","Type":"ContainerStarted","Data":"1e9ffcf6d18abb798ced1e2c030e559effe46dac942f09b4d772146209b1eacd"} Nov 27 10:35:42 crc kubenswrapper[4806]: E1127 10:35:42.954654 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67" podUID="b6163410-3453-43c3-8cd7-5a354e447025" Nov 27 10:35:43 crc kubenswrapper[4806]: E1127 10:35:43.009257 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8" podUID="94f64657-d78c-4f52-8aa1-b23b61d5af68" Nov 27 10:35:43 crc kubenswrapper[4806]: E1127 10:35:43.132716 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb" podUID="856215a1-cdc4-49fb-a522-0c2ca14cd757" Nov 27 10:35:43 crc kubenswrapper[4806]: E1127 10:35:43.325914 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv" podUID="159a9bdd-85be-42dc-8d60-19c3eb1a90e4" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.520939 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm" event={"ID":"4740d40d-863a-40a0-a46f-b515f10e4733","Type":"ContainerStarted","Data":"b30d9f6fd2e3407a3483c327b3574fe42bb6d703e739a7610c4de53ad61f2919"} Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.522514 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.530567 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.543495 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst" event={"ID":"b02aea7d-ec4b-425f-8d13-6205c91ee589","Type":"ContainerStarted","Data":"537fd716282517a196b853ace1a162adf007971d8a0f3a062f66856914e51aae"} Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.544401 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.549594 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.551879 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8" event={"ID":"94f64657-d78c-4f52-8aa1-b23b61d5af68","Type":"ContainerStarted","Data":"a0d7b43856c31764d28c3871d005c5e174d6a6a779bb7bbf408f4c11982bb119"} Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.581593 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" event={"ID":"60db26f8-5bc6-477c-bf21-6842e1b3a9db","Type":"ContainerStarted","Data":"0572206b0abb0028cb99c0fda3040b7a4e5cb4e997670bc6296e1260412cdab4"} Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.585798 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-tdgzm" podStartSLOduration=3.973747763 podStartE2EDuration="42.585722929s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:03.821797369 +0000 UTC m=+808.408388123" lastFinishedPulling="2025-11-27 10:35:42.433772525 +0000 UTC m=+847.020363289" observedRunningTime="2025-11-27 10:35:43.569909879 +0000 UTC m=+848.156500643" watchObservedRunningTime="2025-11-27 10:35:43.585722929 +0000 UTC m=+848.172313713" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.602337 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67" event={"ID":"b6163410-3453-43c3-8cd7-5a354e447025","Type":"ContainerStarted","Data":"6ac6779e19535da540a568f636b3f6c14affbefd38befe790c334e7c7a43e30d"} Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.620042 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb" event={"ID":"856215a1-cdc4-49fb-a522-0c2ca14cd757","Type":"ContainerStarted","Data":"bf2d672b69ca12af0c46789eeb4ae5664895c21ed3c47e5b942f50b6fcaf8ba7"} Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.626574 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-v7tst" podStartSLOduration=3.177934448 podStartE2EDuration="41.626527648s" podCreationTimestamp="2025-11-27 10:35:02 +0000 UTC" firstStartedPulling="2025-11-27 10:35:03.955901399 +0000 UTC m=+808.542492163" lastFinishedPulling="2025-11-27 10:35:42.404494599 +0000 UTC m=+846.991085363" observedRunningTime="2025-11-27 10:35:43.614670397 +0000 UTC m=+848.201261171" watchObservedRunningTime="2025-11-27 10:35:43.626527648 +0000 UTC m=+848.213118412" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.627930 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bz87k" event={"ID":"3104471d-6a15-4454-863e-b5754f748043","Type":"ContainerStarted","Data":"d2da4a1635b1f443c1ddb9671a6736ffcb201656cfb9cb9d6e96b350ed4acbbd"} Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.679320 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" event={"ID":"55fdd641-f8c1-417f-8abb-8adfc95d7008","Type":"ContainerStarted","Data":"8f36074cf9db8e064bd6ddd87ddbdd1897965c8e0730c48162a91a198d44b406"} Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.702281 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj" event={"ID":"609248ce-9b50-48c5-90bd-aa3d2c8e2401","Type":"ContainerStarted","Data":"aa68bd002bc544099eff5de8aac7f969d4796a514a95de41527038db051484f1"} Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.722654 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.737133 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.749824 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv" event={"ID":"159a9bdd-85be-42dc-8d60-19c3eb1a90e4","Type":"ContainerStarted","Data":"e512feb0a19e94ed429f1308188b63c35b5279e47f6dc91efa7be1260c19e5ad"} Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.819527 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.832402 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2" event={"ID":"96cf5347-61b7-4ecb-8093-59b25482cd54","Type":"ContainerStarted","Data":"fff285a670652cc41e6b96254b5bff35a905c7a809c7425664e4436d986507db"} Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.833626 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.860690 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.860766 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.893170 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bz87k" podStartSLOduration=4.006013611 podStartE2EDuration="41.893135113s" podCreationTimestamp="2025-11-27 10:35:02 +0000 UTC" firstStartedPulling="2025-11-27 10:35:04.339335632 +0000 UTC m=+808.925926396" lastFinishedPulling="2025-11-27 10:35:42.226457114 +0000 UTC m=+846.813047898" observedRunningTime="2025-11-27 10:35:43.861151561 +0000 UTC m=+848.447742335" watchObservedRunningTime="2025-11-27 10:35:43.893135113 +0000 UTC m=+848.479725867" Nov 27 10:35:43 crc kubenswrapper[4806]: I1127 10:35:43.939088 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-nkrwj" podStartSLOduration=4.208104257 podStartE2EDuration="42.939069423s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:03.827416085 +0000 UTC m=+808.414006849" lastFinishedPulling="2025-11-27 10:35:42.558381241 +0000 UTC m=+847.144972015" observedRunningTime="2025-11-27 10:35:43.93857581 +0000 UTC m=+848.525166574" watchObservedRunningTime="2025-11-27 10:35:43.939069423 +0000 UTC m=+848.525660177" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.027210 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" podStartSLOduration=3.757990872 podStartE2EDuration="42.027196271s" podCreationTimestamp="2025-11-27 10:35:02 +0000 UTC" firstStartedPulling="2025-11-27 10:35:04.232355577 +0000 UTC m=+808.818946341" lastFinishedPulling="2025-11-27 10:35:42.501560936 +0000 UTC m=+847.088151740" observedRunningTime="2025-11-27 10:35:44.018859519 +0000 UTC m=+848.605450283" watchObservedRunningTime="2025-11-27 10:35:44.027196271 +0000 UTC m=+848.613787035" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.855094 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4" event={"ID":"08fa2215-67c9-4110-b54d-7a18e1641d88","Type":"ContainerStarted","Data":"61838aa3883e9d24e3765cb07e11f164d4e3b5cbf79854fb2fa11f55127ba7c2"} Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.856002 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.864477 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.864878 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6" event={"ID":"33c0104b-4a8a-46fa-bd83-b42f86f96789","Type":"ContainerStarted","Data":"0a09d5838af3482c90f8e57a729af52c24417c4c8c57b420dc3d52e62621a500"} Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.865703 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.872855 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.874728 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26" event={"ID":"5b227bdd-2c69-4053-a116-8085cd157a93","Type":"ContainerStarted","Data":"077c73efbf40c44faa9ac7ab7b72452232600e9879c97de243362cd902c897ed"} Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.875374 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.880351 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m" event={"ID":"4e3d3cf9-5e6d-4963-a2b2-731f4fb5ecb2","Type":"ContainerStarted","Data":"31f8ec2ccb43b3255546903bae1ac4f1ebda7e1863601638b2eec6d2584e121d"} Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.880643 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.890042 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" event={"ID":"f77adb32-dfac-48de-b4a8-c9263cf5729d","Type":"ContainerStarted","Data":"c15bf9c733aa83eaed5685db9c7a6fa07df1c27838fbe8568f892a52ea7f15f6"} Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.890478 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.891671 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" event={"ID":"60db26f8-5bc6-477c-bf21-6842e1b3a9db","Type":"ContainerStarted","Data":"d451a3ba9c8c5120051a073c4d581b7ff1e5867eee1a6687d9e7ffb43546f61e"} Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.891845 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.906461 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.909422 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.909582 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" event={"ID":"d1ceaced-2100-4952-9bc8-cbd5c82886c5","Type":"ContainerStarted","Data":"a969397a5fc1f231d2d9966525294646862902081ef06d9fe3f97c7963009d97"} Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.910546 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.912791 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67" event={"ID":"b6163410-3453-43c3-8cd7-5a354e447025","Type":"ContainerStarted","Data":"593cf9cc03fc0db18ddec6c4e068d97a55e85b61d3daee3156115411683f738d"} Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.913846 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.914073 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.915262 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-hrbw7" event={"ID":"92d56580-f256-46f7-81fa-e370048b4d0f","Type":"ContainerStarted","Data":"88dcf8944f976398bbf6aac45024af6d8221a0fc79e18911ae46a8ba219ac8a0"} Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.920649 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.923686 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t" event={"ID":"671409ea-bf88-4550-a3f7-1fb63a3e89ed","Type":"ContainerStarted","Data":"4f36552a512491c919cc627f06fe59c3f20449de5332ac5c87cc15c1b4a7f459"} Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.924028 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.927329 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.931439 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-6fkj4" podStartSLOduration=4.498978127 podStartE2EDuration="42.931415427s" podCreationTimestamp="2025-11-27 10:35:02 +0000 UTC" firstStartedPulling="2025-11-27 10:35:04.199603835 +0000 UTC m=+808.786194589" lastFinishedPulling="2025-11-27 10:35:42.632041115 +0000 UTC m=+847.218631889" observedRunningTime="2025-11-27 10:35:44.921193762 +0000 UTC m=+849.507784526" watchObservedRunningTime="2025-11-27 10:35:44.931415427 +0000 UTC m=+849.518006191" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.934539 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-kn7j2" podStartSLOduration=4.85975002 podStartE2EDuration="43.934531464s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:03.332220086 +0000 UTC m=+807.918810850" lastFinishedPulling="2025-11-27 10:35:42.40700153 +0000 UTC m=+846.993592294" observedRunningTime="2025-11-27 10:35:44.090617949 +0000 UTC m=+848.677208713" watchObservedRunningTime="2025-11-27 10:35:44.934531464 +0000 UTC m=+849.521122228" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.966925 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" event={"ID":"e5d637d7-a046-4026-a157-abcb8d99ba3b","Type":"ContainerStarted","Data":"4285439ff0f9d3cff1df586aa679b98092821d64f818303b4ab34a0ceab34cb9"} Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.968078 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.975447 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.985155 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-lcsw6" podStartSLOduration=4.903318966 podStartE2EDuration="43.985135725s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:03.479994237 +0000 UTC m=+808.066585001" lastFinishedPulling="2025-11-27 10:35:42.561810986 +0000 UTC m=+847.148401760" observedRunningTime="2025-11-27 10:35:44.983755006 +0000 UTC m=+849.570345770" watchObservedRunningTime="2025-11-27 10:35:44.985135725 +0000 UTC m=+849.571726489" Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.990870 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" event={"ID":"55fdd641-f8c1-417f-8abb-8adfc95d7008","Type":"ContainerStarted","Data":"5cfcad47e5a6fca55040c0263575fe4af8c352f1d330de9fecd0aa66e7fd3b9c"} Nov 27 10:35:44 crc kubenswrapper[4806]: I1127 10:35:44.991143 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.014636 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-ct9ps" event={"ID":"93cca2e7-6982-42d3-8848-10e4f7a70509","Type":"ContainerStarted","Data":"cbfa750efd401a25754baf4179467c571fe379549f06b666d412e72cfaec957b"} Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.017153 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-955677c94-ct9ps" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.017963 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67" podStartSLOduration=3.387823864 podStartE2EDuration="44.01794364s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:03.515752194 +0000 UTC m=+808.102342958" lastFinishedPulling="2025-11-27 10:35:44.14587197 +0000 UTC m=+848.732462734" observedRunningTime="2025-11-27 10:35:45.015519373 +0000 UTC m=+849.602110137" watchObservedRunningTime="2025-11-27 10:35:45.01794364 +0000 UTC m=+849.604534404" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.022982 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-955677c94-ct9ps" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.047318 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d77b94747-xs5lm" podStartSLOduration=4.722780059 podStartE2EDuration="43.047295218s" podCreationTimestamp="2025-11-27 10:35:02 +0000 UTC" firstStartedPulling="2025-11-27 10:35:04.232868513 +0000 UTC m=+808.819459277" lastFinishedPulling="2025-11-27 10:35:42.557383652 +0000 UTC m=+847.143974436" observedRunningTime="2025-11-27 10:35:45.040834339 +0000 UTC m=+849.627425103" watchObservedRunningTime="2025-11-27 10:35:45.047295218 +0000 UTC m=+849.633885982" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.064266 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" event={"ID":"35e9f707-4424-43fc-8d97-bb804524a180","Type":"ContainerStarted","Data":"32c6e03c5e89427d3145996fb7d1076a7e7fca47d2298eaed223f27a4ee5376e"} Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.064315 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.077552 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-v8r26" podStartSLOduration=5.002702458 podStartE2EDuration="44.077527002s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:03.360065002 +0000 UTC m=+807.946655766" lastFinishedPulling="2025-11-27 10:35:42.434889536 +0000 UTC m=+847.021480310" observedRunningTime="2025-11-27 10:35:45.068680365 +0000 UTC m=+849.655271129" watchObservedRunningTime="2025-11-27 10:35:45.077527002 +0000 UTC m=+849.664117766" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.127245 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-4vz5m" podStartSLOduration=5.49907287 podStartE2EDuration="44.127206787s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:03.777225556 +0000 UTC m=+808.363816320" lastFinishedPulling="2025-11-27 10:35:42.405359453 +0000 UTC m=+846.991950237" observedRunningTime="2025-11-27 10:35:45.119604635 +0000 UTC m=+849.706195409" watchObservedRunningTime="2025-11-27 10:35:45.127206787 +0000 UTC m=+849.713797551" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.166936 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" podStartSLOduration=38.049050867 podStartE2EDuration="44.166919315s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:36.052476791 +0000 UTC m=+840.639067575" lastFinishedPulling="2025-11-27 10:35:42.170345249 +0000 UTC m=+846.756936023" observedRunningTime="2025-11-27 10:35:45.161764261 +0000 UTC m=+849.748355015" watchObservedRunningTime="2025-11-27 10:35:45.166919315 +0000 UTC m=+849.753510089" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.225827 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fjhns" podStartSLOduration=4.909586348 podStartE2EDuration="43.225808247s" podCreationTimestamp="2025-11-27 10:35:02 +0000 UTC" firstStartedPulling="2025-11-27 10:35:04.239179098 +0000 UTC m=+808.825769862" lastFinishedPulling="2025-11-27 10:35:42.555400977 +0000 UTC m=+847.141991761" observedRunningTime="2025-11-27 10:35:45.221897127 +0000 UTC m=+849.808487891" watchObservedRunningTime="2025-11-27 10:35:45.225808247 +0000 UTC m=+849.812399011" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.265583 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-955677c94-ct9ps" podStartSLOduration=5.101743648 podStartE2EDuration="44.265562635s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:03.39477579 +0000 UTC m=+807.981366554" lastFinishedPulling="2025-11-27 10:35:42.558594767 +0000 UTC m=+847.145185541" observedRunningTime="2025-11-27 10:35:45.264664701 +0000 UTC m=+849.851255465" watchObservedRunningTime="2025-11-27 10:35:45.265562635 +0000 UTC m=+849.852153399" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.305722 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-j2x9t" podStartSLOduration=5.857528665 podStartE2EDuration="44.305701034s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:04.008703881 +0000 UTC m=+808.595294645" lastFinishedPulling="2025-11-27 10:35:42.45687625 +0000 UTC m=+847.043467014" observedRunningTime="2025-11-27 10:35:45.299638715 +0000 UTC m=+849.886229479" watchObservedRunningTime="2025-11-27 10:35:45.305701034 +0000 UTC m=+849.892291798" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.408784 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-hpmjm" podStartSLOduration=5.946662181 podStartE2EDuration="44.408763339s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:04.049989282 +0000 UTC m=+808.636580046" lastFinishedPulling="2025-11-27 10:35:42.51209042 +0000 UTC m=+847.098681204" observedRunningTime="2025-11-27 10:35:45.407760961 +0000 UTC m=+849.994351745" watchObservedRunningTime="2025-11-27 10:35:45.408763339 +0000 UTC m=+849.995354103" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.411704 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" podStartSLOduration=38.263561107 podStartE2EDuration="44.41169695s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:36.067013336 +0000 UTC m=+840.653604140" lastFinishedPulling="2025-11-27 10:35:42.215149199 +0000 UTC m=+846.801739983" observedRunningTime="2025-11-27 10:35:45.370117111 +0000 UTC m=+849.956707875" watchObservedRunningTime="2025-11-27 10:35:45.41169695 +0000 UTC m=+849.998287714" Nov 27 10:35:45 crc kubenswrapper[4806]: I1127 10:35:45.512903 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" podStartSLOduration=12.16278967 podStartE2EDuration="44.512871662s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:04.050423675 +0000 UTC m=+808.637014439" lastFinishedPulling="2025-11-27 10:35:36.400505666 +0000 UTC m=+840.987096431" observedRunningTime="2025-11-27 10:35:45.481783855 +0000 UTC m=+850.068374619" watchObservedRunningTime="2025-11-27 10:35:45.512871662 +0000 UTC m=+850.099462426" Nov 27 10:35:46 crc kubenswrapper[4806]: I1127 10:35:46.073928 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv" event={"ID":"159a9bdd-85be-42dc-8d60-19c3eb1a90e4","Type":"ContainerStarted","Data":"8a9f6314638c9f58f6e9089163ad2911b68ad8f5db019505d1a8b11def7557e3"} Nov 27 10:35:46 crc kubenswrapper[4806]: I1127 10:35:46.074353 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv" Nov 27 10:35:46 crc kubenswrapper[4806]: I1127 10:35:46.076052 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb" event={"ID":"856215a1-cdc4-49fb-a522-0c2ca14cd757","Type":"ContainerStarted","Data":"48176afa4ef31c74a603a4a9e3aa6641ddbf440dc4ae86b85887427d8ff596bb"} Nov 27 10:35:46 crc kubenswrapper[4806]: I1127 10:35:46.076217 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb" Nov 27 10:35:46 crc kubenswrapper[4806]: I1127 10:35:46.078326 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8" event={"ID":"94f64657-d78c-4f52-8aa1-b23b61d5af68","Type":"ContainerStarted","Data":"453aff5bbccbfb08543e4c5d64c1af8658c3bf2d998cace2c6bdc6926f4c6293"} Nov 27 10:35:46 crc kubenswrapper[4806]: I1127 10:35:46.098043 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv" podStartSLOduration=4.389767745 podStartE2EDuration="45.09802088s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:03.90501536 +0000 UTC m=+808.491606124" lastFinishedPulling="2025-11-27 10:35:44.613268495 +0000 UTC m=+849.199859259" observedRunningTime="2025-11-27 10:35:46.09586533 +0000 UTC m=+850.682456094" watchObservedRunningTime="2025-11-27 10:35:46.09802088 +0000 UTC m=+850.684611644" Nov 27 10:35:46 crc kubenswrapper[4806]: I1127 10:35:46.121972 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8" podStartSLOduration=3.947595224 podStartE2EDuration="45.121941177s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:03.453814887 +0000 UTC m=+808.040405651" lastFinishedPulling="2025-11-27 10:35:44.62816085 +0000 UTC m=+849.214751604" observedRunningTime="2025-11-27 10:35:46.118069429 +0000 UTC m=+850.704660193" watchObservedRunningTime="2025-11-27 10:35:46.121941177 +0000 UTC m=+850.708531941" Nov 27 10:35:46 crc kubenswrapper[4806]: I1127 10:35:46.159162 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb" podStartSLOduration=4.786523678 podStartE2EDuration="45.159140704s" podCreationTimestamp="2025-11-27 10:35:01 +0000 UTC" firstStartedPulling="2025-11-27 10:35:03.963176121 +0000 UTC m=+808.549766885" lastFinishedPulling="2025-11-27 10:35:44.335793147 +0000 UTC m=+848.922383911" observedRunningTime="2025-11-27 10:35:46.154357971 +0000 UTC m=+850.740948735" watchObservedRunningTime="2025-11-27 10:35:46.159140704 +0000 UTC m=+850.745731458" Nov 27 10:35:47 crc kubenswrapper[4806]: I1127 10:35:47.088019 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8" Nov 27 10:35:47 crc kubenswrapper[4806]: I1127 10:35:47.753722 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-55nb9" Nov 27 10:35:48 crc kubenswrapper[4806]: I1127 10:35:48.602863 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh" Nov 27 10:35:48 crc kubenswrapper[4806]: I1127 10:35:48.978652 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7fbdc78c67-7th5m" Nov 27 10:35:51 crc kubenswrapper[4806]: I1127 10:35:51.907548 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-czdn8" Nov 27 10:35:52 crc kubenswrapper[4806]: I1127 10:35:52.144561 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-9jh67" Nov 27 10:35:52 crc kubenswrapper[4806]: I1127 10:35:52.310883 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-m6mcb" Nov 27 10:35:52 crc kubenswrapper[4806]: I1127 10:35:52.573888 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-jtsbv" Nov 27 10:35:52 crc kubenswrapper[4806]: I1127 10:35:52.719567 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-82qfv" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.508756 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-sxjn9"] Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.517292 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-sxjn9" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.521558 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.521558 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.521701 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.521772 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-sxjn9"] Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.536840 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-8sfvx" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.562067 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d-config\") pod \"dnsmasq-dns-675f4bcbfc-sxjn9\" (UID: \"9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-sxjn9" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.562197 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m55wc\" (UniqueName: \"kubernetes.io/projected/9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d-kube-api-access-m55wc\") pod \"dnsmasq-dns-675f4bcbfc-sxjn9\" (UID: \"9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-sxjn9" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.610729 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-7j4v2"] Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.614473 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.618608 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.643292 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-7j4v2"] Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.663203 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m55wc\" (UniqueName: \"kubernetes.io/projected/9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d-kube-api-access-m55wc\") pod \"dnsmasq-dns-675f4bcbfc-sxjn9\" (UID: \"9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-sxjn9" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.663305 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d-config\") pod \"dnsmasq-dns-675f4bcbfc-sxjn9\" (UID: \"9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-sxjn9" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.664358 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d-config\") pod \"dnsmasq-dns-675f4bcbfc-sxjn9\" (UID: \"9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-sxjn9" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.690576 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m55wc\" (UniqueName: \"kubernetes.io/projected/9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d-kube-api-access-m55wc\") pod \"dnsmasq-dns-675f4bcbfc-sxjn9\" (UID: \"9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-sxjn9" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.764577 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048781d0-591e-42b3-9939-c64849bc0b9f-config\") pod \"dnsmasq-dns-78dd6ddcc-7j4v2\" (UID: \"048781d0-591e-42b3-9939-c64849bc0b9f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.765009 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048781d0-591e-42b3-9939-c64849bc0b9f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-7j4v2\" (UID: \"048781d0-591e-42b3-9939-c64849bc0b9f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.765351 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5m5b\" (UniqueName: \"kubernetes.io/projected/048781d0-591e-42b3-9939-c64849bc0b9f-kube-api-access-x5m5b\") pod \"dnsmasq-dns-78dd6ddcc-7j4v2\" (UID: \"048781d0-591e-42b3-9939-c64849bc0b9f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.835101 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-sxjn9" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.866633 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048781d0-591e-42b3-9939-c64849bc0b9f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-7j4v2\" (UID: \"048781d0-591e-42b3-9939-c64849bc0b9f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.866818 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5m5b\" (UniqueName: \"kubernetes.io/projected/048781d0-591e-42b3-9939-c64849bc0b9f-kube-api-access-x5m5b\") pod \"dnsmasq-dns-78dd6ddcc-7j4v2\" (UID: \"048781d0-591e-42b3-9939-c64849bc0b9f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.866863 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048781d0-591e-42b3-9939-c64849bc0b9f-config\") pod \"dnsmasq-dns-78dd6ddcc-7j4v2\" (UID: \"048781d0-591e-42b3-9939-c64849bc0b9f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.867542 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048781d0-591e-42b3-9939-c64849bc0b9f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-7j4v2\" (UID: \"048781d0-591e-42b3-9939-c64849bc0b9f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.867543 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048781d0-591e-42b3-9939-c64849bc0b9f-config\") pod \"dnsmasq-dns-78dd6ddcc-7j4v2\" (UID: \"048781d0-591e-42b3-9939-c64849bc0b9f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.887629 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5m5b\" (UniqueName: \"kubernetes.io/projected/048781d0-591e-42b3-9939-c64849bc0b9f-kube-api-access-x5m5b\") pod \"dnsmasq-dns-78dd6ddcc-7j4v2\" (UID: \"048781d0-591e-42b3-9939-c64849bc0b9f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" Nov 27 10:36:06 crc kubenswrapper[4806]: I1127 10:36:06.939171 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" Nov 27 10:36:07 crc kubenswrapper[4806]: I1127 10:36:07.303250 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-7j4v2"] Nov 27 10:36:07 crc kubenswrapper[4806]: W1127 10:36:07.305535 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod048781d0_591e_42b3_9939_c64849bc0b9f.slice/crio-42a5067a2e7abe0ba6fdc07a5c52f092967d9ee1dcbee9a323e5a22ffcb3438b WatchSource:0}: Error finding container 42a5067a2e7abe0ba6fdc07a5c52f092967d9ee1dcbee9a323e5a22ffcb3438b: Status 404 returned error can't find the container with id 42a5067a2e7abe0ba6fdc07a5c52f092967d9ee1dcbee9a323e5a22ffcb3438b Nov 27 10:36:07 crc kubenswrapper[4806]: I1127 10:36:07.311817 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 10:36:07 crc kubenswrapper[4806]: I1127 10:36:07.416684 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-sxjn9"] Nov 27 10:36:07 crc kubenswrapper[4806]: W1127 10:36:07.428684 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cc42bff_d7bc_46c5_b098_6a3d0b86ba9d.slice/crio-72f4ab063ce67a08a60fbb64bfb228f4b3f3c72b3233d54c9a6e0496b31b92fd WatchSource:0}: Error finding container 72f4ab063ce67a08a60fbb64bfb228f4b3f3c72b3233d54c9a6e0496b31b92fd: Status 404 returned error can't find the container with id 72f4ab063ce67a08a60fbb64bfb228f4b3f3c72b3233d54c9a6e0496b31b92fd Nov 27 10:36:08 crc kubenswrapper[4806]: I1127 10:36:08.303257 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-sxjn9" event={"ID":"9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d","Type":"ContainerStarted","Data":"72f4ab063ce67a08a60fbb64bfb228f4b3f3c72b3233d54c9a6e0496b31b92fd"} Nov 27 10:36:08 crc kubenswrapper[4806]: I1127 10:36:08.305135 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" event={"ID":"048781d0-591e-42b3-9939-c64849bc0b9f","Type":"ContainerStarted","Data":"42a5067a2e7abe0ba6fdc07a5c52f092967d9ee1dcbee9a323e5a22ffcb3438b"} Nov 27 10:36:09 crc kubenswrapper[4806]: I1127 10:36:09.817491 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-sxjn9"] Nov 27 10:36:09 crc kubenswrapper[4806]: I1127 10:36:09.927283 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-v8mkn"] Nov 27 10:36:09 crc kubenswrapper[4806]: I1127 10:36:09.928590 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:09 crc kubenswrapper[4806]: I1127 10:36:09.965302 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-v8mkn"] Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.032421 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5070f2a-0a09-4ff4-b8f6-5978268548d0-config\") pod \"dnsmasq-dns-666b6646f7-v8mkn\" (UID: \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\") " pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.032534 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5070f2a-0a09-4ff4-b8f6-5978268548d0-dns-svc\") pod \"dnsmasq-dns-666b6646f7-v8mkn\" (UID: \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\") " pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.032570 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvnfk\" (UniqueName: \"kubernetes.io/projected/e5070f2a-0a09-4ff4-b8f6-5978268548d0-kube-api-access-wvnfk\") pod \"dnsmasq-dns-666b6646f7-v8mkn\" (UID: \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\") " pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.137384 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5070f2a-0a09-4ff4-b8f6-5978268548d0-config\") pod \"dnsmasq-dns-666b6646f7-v8mkn\" (UID: \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\") " pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.137459 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5070f2a-0a09-4ff4-b8f6-5978268548d0-dns-svc\") pod \"dnsmasq-dns-666b6646f7-v8mkn\" (UID: \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\") " pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.137499 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvnfk\" (UniqueName: \"kubernetes.io/projected/e5070f2a-0a09-4ff4-b8f6-5978268548d0-kube-api-access-wvnfk\") pod \"dnsmasq-dns-666b6646f7-v8mkn\" (UID: \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\") " pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.140175 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5070f2a-0a09-4ff4-b8f6-5978268548d0-config\") pod \"dnsmasq-dns-666b6646f7-v8mkn\" (UID: \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\") " pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.140314 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5070f2a-0a09-4ff4-b8f6-5978268548d0-dns-svc\") pod \"dnsmasq-dns-666b6646f7-v8mkn\" (UID: \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\") " pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.180563 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvnfk\" (UniqueName: \"kubernetes.io/projected/e5070f2a-0a09-4ff4-b8f6-5978268548d0-kube-api-access-wvnfk\") pod \"dnsmasq-dns-666b6646f7-v8mkn\" (UID: \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\") " pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.261606 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.262441 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-7j4v2"] Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.315367 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p2w2s"] Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.316948 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.357854 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p2w2s"] Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.443200 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdr6s\" (UniqueName: \"kubernetes.io/projected/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-kube-api-access-hdr6s\") pod \"dnsmasq-dns-57d769cc4f-p2w2s\" (UID: \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\") " pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.443309 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-p2w2s\" (UID: \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\") " pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.443385 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-config\") pod \"dnsmasq-dns-57d769cc4f-p2w2s\" (UID: \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\") " pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.545446 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-config\") pod \"dnsmasq-dns-57d769cc4f-p2w2s\" (UID: \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\") " pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.545529 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdr6s\" (UniqueName: \"kubernetes.io/projected/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-kube-api-access-hdr6s\") pod \"dnsmasq-dns-57d769cc4f-p2w2s\" (UID: \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\") " pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.545560 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-p2w2s\" (UID: \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\") " pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.546567 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-p2w2s\" (UID: \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\") " pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.547387 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-config\") pod \"dnsmasq-dns-57d769cc4f-p2w2s\" (UID: \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\") " pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.588875 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdr6s\" (UniqueName: \"kubernetes.io/projected/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-kube-api-access-hdr6s\") pod \"dnsmasq-dns-57d769cc4f-p2w2s\" (UID: \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\") " pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:36:10 crc kubenswrapper[4806]: I1127 10:36:10.694865 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:10.999886 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-v8mkn"] Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.044928 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.046316 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.052217 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.052474 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.052629 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.052857 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.053056 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.053423 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.053690 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-rhfhd" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.073456 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.163188 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-config-data\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.163259 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/08137128-e0d1-49d2-abf1-1242c1cb58ce-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.163291 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.163309 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.163334 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.163371 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.163409 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/08137128-e0d1-49d2-abf1-1242c1cb58ce-pod-info\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.163432 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.163447 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-server-conf\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.163496 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.163520 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ncnr\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-kube-api-access-7ncnr\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.269479 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-config-data\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.270596 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/08137128-e0d1-49d2-abf1-1242c1cb58ce-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.270736 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.270762 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.270884 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.270920 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.271092 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/08137128-e0d1-49d2-abf1-1242c1cb58ce-pod-info\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.271294 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.271202 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.271355 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-server-conf\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.271387 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.271625 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ncnr\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-kube-api-access-7ncnr\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.271753 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.272050 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.272712 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-config-data\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.273680 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.274019 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-server-conf\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.281125 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/08137128-e0d1-49d2-abf1-1242c1cb58ce-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.283486 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.310034 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/08137128-e0d1-49d2-abf1-1242c1cb58ce-pod-info\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.311283 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.322782 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ncnr\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-kube-api-access-7ncnr\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.384147 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.485331 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.487014 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.492882 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.493133 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.493291 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.493509 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.493632 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-dl47t" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.493767 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.494909 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.503585 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.579506 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.579558 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.579592 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.579612 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.579628 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.579644 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnwtp\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-kube-api-access-bnwtp\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.579685 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/954f26d7-2bdc-45e3-a07b-d9924469f0a1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.579702 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.579721 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.579738 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/954f26d7-2bdc-45e3-a07b-d9924469f0a1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.579795 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.678212 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.681406 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.681469 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.681488 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.681524 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.681565 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.681583 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.681611 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnwtp\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-kube-api-access-bnwtp\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.681652 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/954f26d7-2bdc-45e3-a07b-d9924469f0a1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.681670 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.681695 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.681715 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/954f26d7-2bdc-45e3-a07b-d9924469f0a1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.681816 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.681923 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.683218 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.683304 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.684018 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.684241 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.687632 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/954f26d7-2bdc-45e3-a07b-d9924469f0a1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.687922 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.704746 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/954f26d7-2bdc-45e3-a07b-d9924469f0a1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.704901 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.706698 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnwtp\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-kube-api-access-bnwtp\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.730935 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:11 crc kubenswrapper[4806]: I1127 10:36:11.818948 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.563126 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.564581 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.577152 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.579344 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-v9tpc" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.580189 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.580271 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.581007 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.585931 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.706309 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.706517 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b30ea2f1-af76-4a43-842c-577ff4f35a20-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.708470 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b30ea2f1-af76-4a43-842c-577ff4f35a20-config-data-default\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.708515 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b30ea2f1-af76-4a43-842c-577ff4f35a20-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.708589 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc9bp\" (UniqueName: \"kubernetes.io/projected/b30ea2f1-af76-4a43-842c-577ff4f35a20-kube-api-access-tc9bp\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.708628 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30ea2f1-af76-4a43-842c-577ff4f35a20-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.708653 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b30ea2f1-af76-4a43-842c-577ff4f35a20-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.708863 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b30ea2f1-af76-4a43-842c-577ff4f35a20-kolla-config\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.810197 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b30ea2f1-af76-4a43-842c-577ff4f35a20-kolla-config\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.810252 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.810296 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b30ea2f1-af76-4a43-842c-577ff4f35a20-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.810314 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b30ea2f1-af76-4a43-842c-577ff4f35a20-config-data-default\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.810331 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b30ea2f1-af76-4a43-842c-577ff4f35a20-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.810358 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc9bp\" (UniqueName: \"kubernetes.io/projected/b30ea2f1-af76-4a43-842c-577ff4f35a20-kube-api-access-tc9bp\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.810381 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30ea2f1-af76-4a43-842c-577ff4f35a20-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.810400 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b30ea2f1-af76-4a43-842c-577ff4f35a20-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.810789 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b30ea2f1-af76-4a43-842c-577ff4f35a20-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.811301 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b30ea2f1-af76-4a43-842c-577ff4f35a20-kolla-config\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.811475 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.823189 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b30ea2f1-af76-4a43-842c-577ff4f35a20-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.823921 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b30ea2f1-af76-4a43-842c-577ff4f35a20-config-data-default\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.834270 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b30ea2f1-af76-4a43-842c-577ff4f35a20-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.835569 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.847922 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b30ea2f1-af76-4a43-842c-577ff4f35a20-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.860621 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc9bp\" (UniqueName: \"kubernetes.io/projected/b30ea2f1-af76-4a43-842c-577ff4f35a20-kube-api-access-tc9bp\") pod \"openstack-galera-0\" (UID: \"b30ea2f1-af76-4a43-842c-577ff4f35a20\") " pod="openstack/openstack-galera-0" Nov 27 10:36:12 crc kubenswrapper[4806]: I1127 10:36:12.902714 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 27 10:36:13 crc kubenswrapper[4806]: I1127 10:36:13.930839 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 27 10:36:13 crc kubenswrapper[4806]: I1127 10:36:13.932644 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:13 crc kubenswrapper[4806]: I1127 10:36:13.936075 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-hh2jl" Nov 27 10:36:13 crc kubenswrapper[4806]: I1127 10:36:13.937655 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 27 10:36:13 crc kubenswrapper[4806]: I1127 10:36:13.937662 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 27 10:36:13 crc kubenswrapper[4806]: I1127 10:36:13.939197 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 27 10:36:13 crc kubenswrapper[4806]: I1127 10:36:13.945822 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.027778 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/46cb1b0e-a6ab-467a-9f49-f8938c164253-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.027927 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.027998 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/46cb1b0e-a6ab-467a-9f49-f8938c164253-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.028073 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/46cb1b0e-a6ab-467a-9f49-f8938c164253-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.028106 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/46cb1b0e-a6ab-467a-9f49-f8938c164253-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.028130 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46cb1b0e-a6ab-467a-9f49-f8938c164253-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.028158 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8hd7\" (UniqueName: \"kubernetes.io/projected/46cb1b0e-a6ab-467a-9f49-f8938c164253-kube-api-access-z8hd7\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.028191 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cb1b0e-a6ab-467a-9f49-f8938c164253-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.129837 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8hd7\" (UniqueName: \"kubernetes.io/projected/46cb1b0e-a6ab-467a-9f49-f8938c164253-kube-api-access-z8hd7\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.129924 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cb1b0e-a6ab-467a-9f49-f8938c164253-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.129987 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/46cb1b0e-a6ab-467a-9f49-f8938c164253-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.130021 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.130064 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/46cb1b0e-a6ab-467a-9f49-f8938c164253-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.130103 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/46cb1b0e-a6ab-467a-9f49-f8938c164253-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.130133 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/46cb1b0e-a6ab-467a-9f49-f8938c164253-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.130155 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46cb1b0e-a6ab-467a-9f49-f8938c164253-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.130502 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.130938 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/46cb1b0e-a6ab-467a-9f49-f8938c164253-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.131458 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/46cb1b0e-a6ab-467a-9f49-f8938c164253-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.131591 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/46cb1b0e-a6ab-467a-9f49-f8938c164253-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.132015 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46cb1b0e-a6ab-467a-9f49-f8938c164253-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.158175 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cb1b0e-a6ab-467a-9f49-f8938c164253-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.160437 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/46cb1b0e-a6ab-467a-9f49-f8938c164253-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.175067 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8hd7\" (UniqueName: \"kubernetes.io/projected/46cb1b0e-a6ab-467a-9f49-f8938c164253-kube-api-access-z8hd7\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.218835 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"46cb1b0e-a6ab-467a-9f49-f8938c164253\") " pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.263117 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.380021 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.381255 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.384956 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.385986 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.386186 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-f7r5g" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.403732 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.434970 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m7bp\" (UniqueName: \"kubernetes.io/projected/d9a162dd-03ba-4444-b03c-bf34c535713a-kube-api-access-7m7bp\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.435057 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9a162dd-03ba-4444-b03c-bf34c535713a-config-data\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.435171 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9a162dd-03ba-4444-b03c-bf34c535713a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.435214 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9a162dd-03ba-4444-b03c-bf34c535713a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.435280 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d9a162dd-03ba-4444-b03c-bf34c535713a-kolla-config\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.450444 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.450516 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.537104 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9a162dd-03ba-4444-b03c-bf34c535713a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.537157 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9a162dd-03ba-4444-b03c-bf34c535713a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.537241 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d9a162dd-03ba-4444-b03c-bf34c535713a-kolla-config\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.537305 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m7bp\" (UniqueName: \"kubernetes.io/projected/d9a162dd-03ba-4444-b03c-bf34c535713a-kube-api-access-7m7bp\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.537328 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9a162dd-03ba-4444-b03c-bf34c535713a-config-data\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.538566 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d9a162dd-03ba-4444-b03c-bf34c535713a-kolla-config\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.539069 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9a162dd-03ba-4444-b03c-bf34c535713a-config-data\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.541989 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9a162dd-03ba-4444-b03c-bf34c535713a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.555692 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m7bp\" (UniqueName: \"kubernetes.io/projected/d9a162dd-03ba-4444-b03c-bf34c535713a-kube-api-access-7m7bp\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.558827 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9a162dd-03ba-4444-b03c-bf34c535713a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d9a162dd-03ba-4444-b03c-bf34c535713a\") " pod="openstack/memcached-0" Nov 27 10:36:14 crc kubenswrapper[4806]: I1127 10:36:14.708677 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 27 10:36:16 crc kubenswrapper[4806]: I1127 10:36:16.105786 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 10:36:16 crc kubenswrapper[4806]: I1127 10:36:16.106929 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 10:36:16 crc kubenswrapper[4806]: I1127 10:36:16.110636 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-6mq2n" Nov 27 10:36:16 crc kubenswrapper[4806]: I1127 10:36:16.157201 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 10:36:16 crc kubenswrapper[4806]: I1127 10:36:16.163485 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh2rj\" (UniqueName: \"kubernetes.io/projected/beb345b2-15fd-4db1-8ea0-c862b7820fc0-kube-api-access-kh2rj\") pod \"kube-state-metrics-0\" (UID: \"beb345b2-15fd-4db1-8ea0-c862b7820fc0\") " pod="openstack/kube-state-metrics-0" Nov 27 10:36:16 crc kubenswrapper[4806]: I1127 10:36:16.264660 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh2rj\" (UniqueName: \"kubernetes.io/projected/beb345b2-15fd-4db1-8ea0-c862b7820fc0-kube-api-access-kh2rj\") pod \"kube-state-metrics-0\" (UID: \"beb345b2-15fd-4db1-8ea0-c862b7820fc0\") " pod="openstack/kube-state-metrics-0" Nov 27 10:36:16 crc kubenswrapper[4806]: I1127 10:36:16.305079 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh2rj\" (UniqueName: \"kubernetes.io/projected/beb345b2-15fd-4db1-8ea0-c862b7820fc0-kube-api-access-kh2rj\") pod \"kube-state-metrics-0\" (UID: \"beb345b2-15fd-4db1-8ea0-c862b7820fc0\") " pod="openstack/kube-state-metrics-0" Nov 27 10:36:16 crc kubenswrapper[4806]: I1127 10:36:16.418000 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" event={"ID":"e5070f2a-0a09-4ff4-b8f6-5978268548d0","Type":"ContainerStarted","Data":"af935e12933b42c76c2fb28ab0529071a24ecb4b936801377f7b242facf7aadf"} Nov 27 10:36:16 crc kubenswrapper[4806]: I1127 10:36:16.443157 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.829381 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p2w2s"] Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.844170 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-l82sn"] Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.845290 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l82sn" Nov 27 10:36:19 crc kubenswrapper[4806]: W1127 10:36:19.854526 4806 reflector.go:561] object-"openstack"/"ovncontroller-scripts": failed to list *v1.ConfigMap: configmaps "ovncontroller-scripts" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Nov 27 10:36:19 crc kubenswrapper[4806]: E1127 10:36:19.854601 4806 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"ovncontroller-scripts\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovncontroller-scripts\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 27 10:36:19 crc kubenswrapper[4806]: W1127 10:36:19.857668 4806 reflector.go:561] object-"openstack"/"ovncontroller-ovncontroller-dockercfg-267bj": failed to list *v1.Secret: secrets "ovncontroller-ovncontroller-dockercfg-267bj" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Nov 27 10:36:19 crc kubenswrapper[4806]: E1127 10:36:19.857721 4806 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"ovncontroller-ovncontroller-dockercfg-267bj\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovncontroller-ovncontroller-dockercfg-267bj\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 27 10:36:19 crc kubenswrapper[4806]: W1127 10:36:19.857763 4806 reflector.go:561] object-"openstack"/"cert-ovncontroller-ovndbs": failed to list *v1.Secret: secrets "cert-ovncontroller-ovndbs" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Nov 27 10:36:19 crc kubenswrapper[4806]: E1127 10:36:19.857773 4806 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"cert-ovncontroller-ovndbs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-ovncontroller-ovndbs\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.868473 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-f769m"] Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.878411 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.902288 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-f769m"] Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.918360 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l82sn"] Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.933959 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d814f280-e3ac-426c-bafa-4a4ed3f51320-etc-ovs\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.934581 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2cb892e8-358b-477b-8741-4211a599bf6c-scripts\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.934893 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl54w\" (UniqueName: \"kubernetes.io/projected/2cb892e8-358b-477b-8741-4211a599bf6c-kube-api-access-gl54w\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.935058 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d814f280-e3ac-426c-bafa-4a4ed3f51320-scripts\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.940721 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d814f280-e3ac-426c-bafa-4a4ed3f51320-var-log\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.940930 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2cb892e8-358b-477b-8741-4211a599bf6c-var-run-ovn\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.941043 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d814f280-e3ac-426c-bafa-4a4ed3f51320-var-run\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.941164 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2cb892e8-358b-477b-8741-4211a599bf6c-var-run\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.941286 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2cb892e8-358b-477b-8741-4211a599bf6c-var-log-ovn\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.941371 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cb892e8-358b-477b-8741-4211a599bf6c-ovn-controller-tls-certs\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.941500 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d814f280-e3ac-426c-bafa-4a4ed3f51320-var-lib\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.941659 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smwwq\" (UniqueName: \"kubernetes.io/projected/d814f280-e3ac-426c-bafa-4a4ed3f51320-kube-api-access-smwwq\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:19 crc kubenswrapper[4806]: I1127 10:36:19.941733 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cb892e8-358b-477b-8741-4211a599bf6c-combined-ca-bundle\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.042516 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2cb892e8-358b-477b-8741-4211a599bf6c-var-run\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.042794 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2cb892e8-358b-477b-8741-4211a599bf6c-var-log-ovn\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.042816 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cb892e8-358b-477b-8741-4211a599bf6c-ovn-controller-tls-certs\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.042836 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d814f280-e3ac-426c-bafa-4a4ed3f51320-var-lib\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.042854 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smwwq\" (UniqueName: \"kubernetes.io/projected/d814f280-e3ac-426c-bafa-4a4ed3f51320-kube-api-access-smwwq\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.042872 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cb892e8-358b-477b-8741-4211a599bf6c-combined-ca-bundle\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.042893 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d814f280-e3ac-426c-bafa-4a4ed3f51320-etc-ovs\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.042921 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2cb892e8-358b-477b-8741-4211a599bf6c-scripts\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.042955 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl54w\" (UniqueName: \"kubernetes.io/projected/2cb892e8-358b-477b-8741-4211a599bf6c-kube-api-access-gl54w\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.042974 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d814f280-e3ac-426c-bafa-4a4ed3f51320-scripts\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.042991 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d814f280-e3ac-426c-bafa-4a4ed3f51320-var-log\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.043020 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2cb892e8-358b-477b-8741-4211a599bf6c-var-run-ovn\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.043038 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d814f280-e3ac-426c-bafa-4a4ed3f51320-var-run\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.043056 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2cb892e8-358b-477b-8741-4211a599bf6c-var-run\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.043119 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d814f280-e3ac-426c-bafa-4a4ed3f51320-var-run\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.043186 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d814f280-e3ac-426c-bafa-4a4ed3f51320-etc-ovs\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.043317 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2cb892e8-358b-477b-8741-4211a599bf6c-var-log-ovn\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.043502 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d814f280-e3ac-426c-bafa-4a4ed3f51320-var-lib\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.043544 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d814f280-e3ac-426c-bafa-4a4ed3f51320-var-log\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.043587 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2cb892e8-358b-477b-8741-4211a599bf6c-var-run-ovn\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.051040 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cb892e8-358b-477b-8741-4211a599bf6c-combined-ca-bundle\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.062172 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl54w\" (UniqueName: \"kubernetes.io/projected/2cb892e8-358b-477b-8741-4211a599bf6c-kube-api-access-gl54w\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.062607 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smwwq\" (UniqueName: \"kubernetes.io/projected/d814f280-e3ac-426c-bafa-4a4ed3f51320-kube-api-access-smwwq\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.854455 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.857369 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2cb892e8-358b-477b-8741-4211a599bf6c-scripts\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:20 crc kubenswrapper[4806]: I1127 10:36:20.869697 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d814f280-e3ac-426c-bafa-4a4ed3f51320-scripts\") pod \"ovn-controller-ovs-f769m\" (UID: \"d814f280-e3ac-426c-bafa-4a4ed3f51320\") " pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:21 crc kubenswrapper[4806]: I1127 10:36:21.026582 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-267bj" Nov 27 10:36:21 crc kubenswrapper[4806]: E1127 10:36:21.044261 4806 secret.go:188] Couldn't get secret openstack/cert-ovncontroller-ovndbs: failed to sync secret cache: timed out waiting for the condition Nov 27 10:36:21 crc kubenswrapper[4806]: E1127 10:36:21.044390 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2cb892e8-358b-477b-8741-4211a599bf6c-ovn-controller-tls-certs podName:2cb892e8-358b-477b-8741-4211a599bf6c nodeName:}" failed. No retries permitted until 2025-11-27 10:36:21.544358874 +0000 UTC m=+886.130949638 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovn-controller-tls-certs" (UniqueName: "kubernetes.io/secret/2cb892e8-358b-477b-8741-4211a599bf6c-ovn-controller-tls-certs") pod "ovn-controller-l82sn" (UID: "2cb892e8-358b-477b-8741-4211a599bf6c") : failed to sync secret cache: timed out waiting for the condition Nov 27 10:36:21 crc kubenswrapper[4806]: I1127 10:36:21.109219 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:36:21 crc kubenswrapper[4806]: I1127 10:36:21.453605 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 27 10:36:21 crc kubenswrapper[4806]: I1127 10:36:21.573833 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cb892e8-358b-477b-8741-4211a599bf6c-ovn-controller-tls-certs\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:21 crc kubenswrapper[4806]: I1127 10:36:21.580178 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cb892e8-358b-477b-8741-4211a599bf6c-ovn-controller-tls-certs\") pod \"ovn-controller-l82sn\" (UID: \"2cb892e8-358b-477b-8741-4211a599bf6c\") " pod="openstack/ovn-controller-l82sn" Nov 27 10:36:21 crc kubenswrapper[4806]: I1127 10:36:21.662054 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l82sn" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.710224 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.712281 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.714879 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.715256 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.715493 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.715657 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.715697 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-h2ns2" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.725418 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.899555 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/08068118-444f-4ea3-9c3c-7f0768a257b2-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.899654 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.899703 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08068118-444f-4ea3-9c3c-7f0768a257b2-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.899768 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkbp7\" (UniqueName: \"kubernetes.io/projected/08068118-444f-4ea3-9c3c-7f0768a257b2-kube-api-access-dkbp7\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.899854 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08068118-444f-4ea3-9c3c-7f0768a257b2-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.899978 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/08068118-444f-4ea3-9c3c-7f0768a257b2-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.900073 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08068118-444f-4ea3-9c3c-7f0768a257b2-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:22 crc kubenswrapper[4806]: I1127 10:36:22.900131 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08068118-444f-4ea3-9c3c-7f0768a257b2-config\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.001894 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08068118-444f-4ea3-9c3c-7f0768a257b2-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.001975 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkbp7\" (UniqueName: \"kubernetes.io/projected/08068118-444f-4ea3-9c3c-7f0768a257b2-kube-api-access-dkbp7\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.002013 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08068118-444f-4ea3-9c3c-7f0768a257b2-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.002057 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/08068118-444f-4ea3-9c3c-7f0768a257b2-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.002091 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08068118-444f-4ea3-9c3c-7f0768a257b2-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.002134 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08068118-444f-4ea3-9c3c-7f0768a257b2-config\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.002178 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/08068118-444f-4ea3-9c3c-7f0768a257b2-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.002247 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.002682 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.006441 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/08068118-444f-4ea3-9c3c-7f0768a257b2-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.008119 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08068118-444f-4ea3-9c3c-7f0768a257b2-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.008177 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08068118-444f-4ea3-9c3c-7f0768a257b2-config\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.012832 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/08068118-444f-4ea3-9c3c-7f0768a257b2-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.014393 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08068118-444f-4ea3-9c3c-7f0768a257b2-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.022406 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08068118-444f-4ea3-9c3c-7f0768a257b2-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.037054 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkbp7\" (UniqueName: \"kubernetes.io/projected/08068118-444f-4ea3-9c3c-7f0768a257b2-kube-api-access-dkbp7\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.041035 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"08068118-444f-4ea3-9c3c-7f0768a257b2\") " pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.068290 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.487778 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.489562 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.496220 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.496402 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.496439 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-sx6nz" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.496447 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.506922 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.620717 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9fcn\" (UniqueName: \"kubernetes.io/projected/d8134ca5-1cd3-4793-875e-8ea7bde9c407-kube-api-access-w9fcn\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.620805 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8134ca5-1cd3-4793-875e-8ea7bde9c407-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.620841 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.620869 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8134ca5-1cd3-4793-875e-8ea7bde9c407-config\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.620900 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8134ca5-1cd3-4793-875e-8ea7bde9c407-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.620950 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8134ca5-1cd3-4793-875e-8ea7bde9c407-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.620971 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8134ca5-1cd3-4793-875e-8ea7bde9c407-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.620999 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d8134ca5-1cd3-4793-875e-8ea7bde9c407-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.723108 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8134ca5-1cd3-4793-875e-8ea7bde9c407-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.723182 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.723245 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8134ca5-1cd3-4793-875e-8ea7bde9c407-config\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.723288 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8134ca5-1cd3-4793-875e-8ea7bde9c407-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.723383 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8134ca5-1cd3-4793-875e-8ea7bde9c407-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.723425 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8134ca5-1cd3-4793-875e-8ea7bde9c407-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.723439 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.723453 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d8134ca5-1cd3-4793-875e-8ea7bde9c407-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.723689 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9fcn\" (UniqueName: \"kubernetes.io/projected/d8134ca5-1cd3-4793-875e-8ea7bde9c407-kube-api-access-w9fcn\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.739567 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d8134ca5-1cd3-4793-875e-8ea7bde9c407-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.740661 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8134ca5-1cd3-4793-875e-8ea7bde9c407-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.740940 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8134ca5-1cd3-4793-875e-8ea7bde9c407-config\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.753085 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8134ca5-1cd3-4793-875e-8ea7bde9c407-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.760775 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8134ca5-1cd3-4793-875e-8ea7bde9c407-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.761157 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8134ca5-1cd3-4793-875e-8ea7bde9c407-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.767051 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9fcn\" (UniqueName: \"kubernetes.io/projected/d8134ca5-1cd3-4793-875e-8ea7bde9c407-kube-api-access-w9fcn\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.815830 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d8134ca5-1cd3-4793-875e-8ea7bde9c407\") " pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:23 crc kubenswrapper[4806]: I1127 10:36:23.823496 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 27 10:36:25 crc kubenswrapper[4806]: W1127 10:36:25.205685 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87df61a9_cd16_4ccb_b8bc_f9fd9cca47d7.slice/crio-858a7d68323cf73642569f5a700d28eee0b02322a63c765b07e0f2c0bb74aefe WatchSource:0}: Error finding container 858a7d68323cf73642569f5a700d28eee0b02322a63c765b07e0f2c0bb74aefe: Status 404 returned error can't find the container with id 858a7d68323cf73642569f5a700d28eee0b02322a63c765b07e0f2c0bb74aefe Nov 27 10:36:25 crc kubenswrapper[4806]: I1127 10:36:25.531950 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" event={"ID":"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7","Type":"ContainerStarted","Data":"858a7d68323cf73642569f5a700d28eee0b02322a63c765b07e0f2c0bb74aefe"} Nov 27 10:36:25 crc kubenswrapper[4806]: I1127 10:36:25.730967 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 27 10:36:25 crc kubenswrapper[4806]: I1127 10:36:25.811842 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 10:36:25 crc kubenswrapper[4806]: I1127 10:36:25.986212 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 27 10:36:25 crc kubenswrapper[4806]: I1127 10:36:25.992273 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 10:36:26 crc kubenswrapper[4806]: W1127 10:36:26.477271 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46cb1b0e_a6ab_467a_9f49_f8938c164253.slice/crio-ff642a686620c478f196f921d54be56a493b6045ccaf9b4dc8461631b1fe6912 WatchSource:0}: Error finding container ff642a686620c478f196f921d54be56a493b6045ccaf9b4dc8461631b1fe6912: Status 404 returned error can't find the container with id ff642a686620c478f196f921d54be56a493b6045ccaf9b4dc8461631b1fe6912 Nov 27 10:36:26 crc kubenswrapper[4806]: E1127 10:36:26.479104 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 27 10:36:26 crc kubenswrapper[4806]: E1127 10:36:26.479354 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x5m5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-7j4v2_openstack(048781d0-591e-42b3-9939-c64849bc0b9f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:36:26 crc kubenswrapper[4806]: E1127 10:36:26.483909 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" podUID="048781d0-591e-42b3-9939-c64849bc0b9f" Nov 27 10:36:26 crc kubenswrapper[4806]: W1127 10:36:26.497350 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb30ea2f1_af76_4a43_842c_577ff4f35a20.slice/crio-599e502f4fd9c6c667be93fe1693ea8e9ce004b10e4efdc73dc49c5b55e1696e WatchSource:0}: Error finding container 599e502f4fd9c6c667be93fe1693ea8e9ce004b10e4efdc73dc49c5b55e1696e: Status 404 returned error can't find the container with id 599e502f4fd9c6c667be93fe1693ea8e9ce004b10e4efdc73dc49c5b55e1696e Nov 27 10:36:26 crc kubenswrapper[4806]: E1127 10:36:26.498203 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 27 10:36:26 crc kubenswrapper[4806]: E1127 10:36:26.498491 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m55wc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-sxjn9_openstack(9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:36:26 crc kubenswrapper[4806]: E1127 10:36:26.499851 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-sxjn9" podUID="9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d" Nov 27 10:36:26 crc kubenswrapper[4806]: I1127 10:36:26.590925 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"08137128-e0d1-49d2-abf1-1242c1cb58ce","Type":"ContainerStarted","Data":"a39aa1d2e145b761ba1c1c263afbca6fea9593d6c18b6dd24b3047a20a4e09fc"} Nov 27 10:36:26 crc kubenswrapper[4806]: I1127 10:36:26.597159 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b30ea2f1-af76-4a43-842c-577ff4f35a20","Type":"ContainerStarted","Data":"599e502f4fd9c6c667be93fe1693ea8e9ce004b10e4efdc73dc49c5b55e1696e"} Nov 27 10:36:26 crc kubenswrapper[4806]: I1127 10:36:26.634623 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"46cb1b0e-a6ab-467a-9f49-f8938c164253","Type":"ContainerStarted","Data":"ff642a686620c478f196f921d54be56a493b6045ccaf9b4dc8461631b1fe6912"} Nov 27 10:36:26 crc kubenswrapper[4806]: I1127 10:36:26.671583 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"954f26d7-2bdc-45e3-a07b-d9924469f0a1","Type":"ContainerStarted","Data":"9ec78ae4e8d5273abfdda51f5cdf42fb72508d56b70ca2134e687b9e5b6f156b"} Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.099214 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.157428 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 27 10:36:27 crc kubenswrapper[4806]: W1127 10:36:27.221587 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbeb345b2_15fd_4db1_8ea0_c862b7820fc0.slice/crio-be3f12a2906eee11239ad63bf0c5aa8de6cb1041aa34c4da893b37649eadd8aa WatchSource:0}: Error finding container be3f12a2906eee11239ad63bf0c5aa8de6cb1041aa34c4da893b37649eadd8aa: Status 404 returned error can't find the container with id be3f12a2906eee11239ad63bf0c5aa8de6cb1041aa34c4da893b37649eadd8aa Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.251173 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.265455 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-sxjn9" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.358255 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048781d0-591e-42b3-9939-c64849bc0b9f-config\") pod \"048781d0-591e-42b3-9939-c64849bc0b9f\" (UID: \"048781d0-591e-42b3-9939-c64849bc0b9f\") " Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.358334 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048781d0-591e-42b3-9939-c64849bc0b9f-dns-svc\") pod \"048781d0-591e-42b3-9939-c64849bc0b9f\" (UID: \"048781d0-591e-42b3-9939-c64849bc0b9f\") " Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.358358 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5m5b\" (UniqueName: \"kubernetes.io/projected/048781d0-591e-42b3-9939-c64849bc0b9f-kube-api-access-x5m5b\") pod \"048781d0-591e-42b3-9939-c64849bc0b9f\" (UID: \"048781d0-591e-42b3-9939-c64849bc0b9f\") " Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.358377 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m55wc\" (UniqueName: \"kubernetes.io/projected/9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d-kube-api-access-m55wc\") pod \"9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d\" (UID: \"9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d\") " Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.358592 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d-config\") pod \"9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d\" (UID: \"9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d\") " Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.359057 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/048781d0-591e-42b3-9939-c64849bc0b9f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "048781d0-591e-42b3-9939-c64849bc0b9f" (UID: "048781d0-591e-42b3-9939-c64849bc0b9f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.359152 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/048781d0-591e-42b3-9939-c64849bc0b9f-config" (OuterVolumeSpecName: "config") pod "048781d0-591e-42b3-9939-c64849bc0b9f" (UID: "048781d0-591e-42b3-9939-c64849bc0b9f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.359466 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d-config" (OuterVolumeSpecName: "config") pod "9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d" (UID: "9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.363982 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d-kube-api-access-m55wc" (OuterVolumeSpecName: "kube-api-access-m55wc") pod "9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d" (UID: "9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d"). InnerVolumeSpecName "kube-api-access-m55wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.364648 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/048781d0-591e-42b3-9939-c64849bc0b9f-kube-api-access-x5m5b" (OuterVolumeSpecName: "kube-api-access-x5m5b") pod "048781d0-591e-42b3-9939-c64849bc0b9f" (UID: "048781d0-591e-42b3-9939-c64849bc0b9f"). InnerVolumeSpecName "kube-api-access-x5m5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.461026 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.461062 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/048781d0-591e-42b3-9939-c64849bc0b9f-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.461075 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5m5b\" (UniqueName: \"kubernetes.io/projected/048781d0-591e-42b3-9939-c64849bc0b9f-kube-api-access-x5m5b\") on node \"crc\" DevicePath \"\"" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.461087 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/048781d0-591e-42b3-9939-c64849bc0b9f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.461096 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m55wc\" (UniqueName: \"kubernetes.io/projected/9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d-kube-api-access-m55wc\") on node \"crc\" DevicePath \"\"" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.600575 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l82sn"] Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.702729 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d9a162dd-03ba-4444-b03c-bf34c535713a","Type":"ContainerStarted","Data":"29baf6a54f0ec0ae640526c30846fbabc7006e5ee86784c221abd9b52f492ba9"} Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.704708 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-sxjn9" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.704760 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-sxjn9" event={"ID":"9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d","Type":"ContainerDied","Data":"72f4ab063ce67a08a60fbb64bfb228f4b3f3c72b3233d54c9a6e0496b31b92fd"} Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.719363 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" event={"ID":"048781d0-591e-42b3-9939-c64849bc0b9f","Type":"ContainerDied","Data":"42a5067a2e7abe0ba6fdc07a5c52f092967d9ee1dcbee9a323e5a22ffcb3438b"} Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.719507 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-7j4v2" Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.723480 4806 generic.go:334] "Generic (PLEG): container finished" podID="87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7" containerID="dfb4fb43d38b4b4a7ccfd0dde865ffa380a9bf9fe33fe75528ba0fe71a12ab64" exitCode=0 Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.723545 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" event={"ID":"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7","Type":"ContainerDied","Data":"dfb4fb43d38b4b4a7ccfd0dde865ffa380a9bf9fe33fe75528ba0fe71a12ab64"} Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.729294 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"beb345b2-15fd-4db1-8ea0-c862b7820fc0","Type":"ContainerStarted","Data":"be3f12a2906eee11239ad63bf0c5aa8de6cb1041aa34c4da893b37649eadd8aa"} Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.743402 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l82sn" event={"ID":"2cb892e8-358b-477b-8741-4211a599bf6c","Type":"ContainerStarted","Data":"b60512c827ef5c8c51fa10d849fb74057678ea1652f7cb41f545c3178c6dac5d"} Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.745105 4806 generic.go:334] "Generic (PLEG): container finished" podID="e5070f2a-0a09-4ff4-b8f6-5978268548d0" containerID="d9077c1972fcc3000136b40d14f8c5cbbd17976f663763ce0eb41fc92791790b" exitCode=0 Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.745133 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" event={"ID":"e5070f2a-0a09-4ff4-b8f6-5978268548d0","Type":"ContainerDied","Data":"d9077c1972fcc3000136b40d14f8c5cbbd17976f663763ce0eb41fc92791790b"} Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.810410 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-sxjn9"] Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.832854 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-sxjn9"] Nov 27 10:36:27 crc kubenswrapper[4806]: I1127 10:36:27.965775 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-7j4v2"] Nov 27 10:36:28 crc kubenswrapper[4806]: I1127 10:36:28.002150 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-7j4v2"] Nov 27 10:36:28 crc kubenswrapper[4806]: I1127 10:36:28.136778 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="048781d0-591e-42b3-9939-c64849bc0b9f" path="/var/lib/kubelet/pods/048781d0-591e-42b3-9939-c64849bc0b9f/volumes" Nov 27 10:36:28 crc kubenswrapper[4806]: I1127 10:36:28.140599 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d" path="/var/lib/kubelet/pods/9cc42bff-d7bc-46c5-b098-6a3d0b86ba9d/volumes" Nov 27 10:36:28 crc kubenswrapper[4806]: I1127 10:36:28.146248 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 27 10:36:28 crc kubenswrapper[4806]: E1127 10:36:28.491516 4806 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 27 10:36:28 crc kubenswrapper[4806]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/e5070f2a-0a09-4ff4-b8f6-5978268548d0/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 27 10:36:28 crc kubenswrapper[4806]: > podSandboxID="af935e12933b42c76c2fb28ab0529071a24ecb4b936801377f7b242facf7aadf" Nov 27 10:36:28 crc kubenswrapper[4806]: E1127 10:36:28.492094 4806 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 27 10:36:28 crc kubenswrapper[4806]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wvnfk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-v8mkn_openstack(e5070f2a-0a09-4ff4-b8f6-5978268548d0): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/e5070f2a-0a09-4ff4-b8f6-5978268548d0/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 27 10:36:28 crc kubenswrapper[4806]: > logger="UnhandledError" Nov 27 10:36:28 crc kubenswrapper[4806]: E1127 10:36:28.493924 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/e5070f2a-0a09-4ff4-b8f6-5978268548d0/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" podUID="e5070f2a-0a09-4ff4-b8f6-5978268548d0" Nov 27 10:36:28 crc kubenswrapper[4806]: I1127 10:36:28.761613 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" event={"ID":"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7","Type":"ContainerStarted","Data":"6d6f1bb081e4d9609ef5fc110a6d05da448158bf2d1c12676d23807fc8dd6094"} Nov 27 10:36:28 crc kubenswrapper[4806]: I1127 10:36:28.763629 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:36:28 crc kubenswrapper[4806]: I1127 10:36:28.765021 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d8134ca5-1cd3-4793-875e-8ea7bde9c407","Type":"ContainerStarted","Data":"4971842516caba34d77249bb5a8f929b2b27eade60a9adfb6c10526608246926"} Nov 27 10:36:28 crc kubenswrapper[4806]: I1127 10:36:28.795828 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" podStartSLOduration=17.306688121 podStartE2EDuration="18.795798887s" podCreationTimestamp="2025-11-27 10:36:10 +0000 UTC" firstStartedPulling="2025-11-27 10:36:25.212512361 +0000 UTC m=+889.799103125" lastFinishedPulling="2025-11-27 10:36:26.701623127 +0000 UTC m=+891.288213891" observedRunningTime="2025-11-27 10:36:28.783993678 +0000 UTC m=+893.370584442" watchObservedRunningTime="2025-11-27 10:36:28.795798887 +0000 UTC m=+893.382389681" Nov 27 10:36:29 crc kubenswrapper[4806]: I1127 10:36:29.044898 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 27 10:36:29 crc kubenswrapper[4806]: I1127 10:36:29.776177 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"08068118-444f-4ea3-9c3c-7f0768a257b2","Type":"ContainerStarted","Data":"cc104b644a6824e6ceda5894d14b19f0becdf19711981f946ff2ea5d301397de"} Nov 27 10:36:29 crc kubenswrapper[4806]: I1127 10:36:29.789332 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-f769m"] Nov 27 10:36:31 crc kubenswrapper[4806]: W1127 10:36:31.194446 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd814f280_e3ac_426c_bafa_4a4ed3f51320.slice/crio-5b51c0bb4a796da948ad4f46c694ba4e02333eb05c908efb787b2844e68ce63c WatchSource:0}: Error finding container 5b51c0bb4a796da948ad4f46c694ba4e02333eb05c908efb787b2844e68ce63c: Status 404 returned error can't find the container with id 5b51c0bb4a796da948ad4f46c694ba4e02333eb05c908efb787b2844e68ce63c Nov 27 10:36:31 crc kubenswrapper[4806]: I1127 10:36:31.811044 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f769m" event={"ID":"d814f280-e3ac-426c-bafa-4a4ed3f51320","Type":"ContainerStarted","Data":"5b51c0bb4a796da948ad4f46c694ba4e02333eb05c908efb787b2844e68ce63c"} Nov 27 10:36:35 crc kubenswrapper[4806]: I1127 10:36:35.697077 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:36:35 crc kubenswrapper[4806]: I1127 10:36:35.777141 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-v8mkn"] Nov 27 10:36:44 crc kubenswrapper[4806]: I1127 10:36:44.450885 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:36:44 crc kubenswrapper[4806]: I1127 10:36:44.451550 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:36:45 crc kubenswrapper[4806]: E1127 10:36:45.479838 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Nov 27 10:36:45 crc kubenswrapper[4806]: E1127 10:36:45.480068 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z8hd7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(46cb1b0e-a6ab-467a-9f49-f8938c164253): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:36:45 crc kubenswrapper[4806]: E1127 10:36:45.481545 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="46cb1b0e-a6ab-467a-9f49-f8938c164253" Nov 27 10:36:45 crc kubenswrapper[4806]: E1127 10:36:45.671821 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified" Nov 27 10:36:45 crc kubenswrapper[4806]: E1127 10:36:45.672059 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:ovsdb-server-init,Image:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,Command:[/usr/local/bin/container-scripts/init-ovsdb-server.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n596h6h565h66bh649hc9h57bh5bbhd4h56fh7dh7fhb4h68dh75h546h75h547h6h66fhf5h55ch559h658h65bhb5hb6h666hdfh54ch545h5f5q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-ovs,ReadOnly:false,MountPath:/etc/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log,ReadOnly:false,MountPath:/var/log/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-lib,ReadOnly:false,MountPath:/var/lib/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-smwwq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-ovs-f769m_openstack(d814f280-e3ac-426c-bafa-4a4ed3f51320): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:36:45 crc kubenswrapper[4806]: E1127 10:36:45.673264 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-ovs-f769m" podUID="d814f280-e3ac-426c-bafa-4a4ed3f51320" Nov 27 10:36:45 crc kubenswrapper[4806]: E1127 10:36:45.750142 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Nov 27 10:36:45 crc kubenswrapper[4806]: E1127 10:36:45.750419 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tc9bp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(b30ea2f1-af76-4a43-842c-577ff4f35a20): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:36:45 crc kubenswrapper[4806]: E1127 10:36:45.751902 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="b30ea2f1-af76-4a43-842c-577ff4f35a20" Nov 27 10:36:45 crc kubenswrapper[4806]: E1127 10:36:45.942630 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="46cb1b0e-a6ab-467a-9f49-f8938c164253" Nov 27 10:36:45 crc kubenswrapper[4806]: E1127 10:36:45.942657 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="b30ea2f1-af76-4a43-842c-577ff4f35a20" Nov 27 10:36:45 crc kubenswrapper[4806]: E1127 10:36:45.942741 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdb-server-init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified\\\"\"" pod="openstack/ovn-controller-ovs-f769m" podUID="d814f280-e3ac-426c-bafa-4a4ed3f51320" Nov 27 10:36:46 crc kubenswrapper[4806]: E1127 10:36:46.212671 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified" Nov 27 10:36:46 crc kubenswrapper[4806]: E1127 10:36:46.214642 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-controller,Image:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,Command:[ovn-controller --pidfile unix:/run/openvswitch/db.sock --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n596h6h565h66bh649hc9h57bh5bbhd4h56fh7dh7fhb4h68dh75h546h75h547h6h66fhf5h55ch559h658h65bhb5hb6h666hdfh54ch545h5f5q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run-ovn,ReadOnly:false,MountPath:/var/run/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log-ovn,ReadOnly:false,MountPath:/var/log/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gl54w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_liveness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_readiness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/share/ovn/scripts/ovn-ctl stop_controller],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-l82sn_openstack(2cb892e8-358b-477b-8741-4211a599bf6c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:36:46 crc kubenswrapper[4806]: E1127 10:36:46.216110 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-controller-l82sn" podUID="2cb892e8-358b-477b-8741-4211a599bf6c" Nov 27 10:36:46 crc kubenswrapper[4806]: E1127 10:36:46.416103 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified" Nov 27 10:36:46 crc kubenswrapper[4806]: E1127 10:36:46.416498 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-sb,Image:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68h657h55bh644h698hf4h6fhdh7h646hddh5c7h667hf6h89h694h5f8h96h5b6hf6h86hdfh576h696h77hdbhcch5f5h645h64fh5d8h55dq,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-sb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w9fcn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(d8134ca5-1cd3-4793-875e-8ea7bde9c407): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:36:46 crc kubenswrapper[4806]: E1127 10:36:46.763208 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified" Nov 27 10:36:46 crc kubenswrapper[4806]: E1127 10:36:46.763535 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n655h8bh75h584h546h56dh65dh5bh66dh54bh54h5fh5bh5dhb9hcdh587hf9h56dh688h549hf4h7dh5bbh5ddh667hbfh5f6h7fhc9h5b8h7dq,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dkbp7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(08068118-444f-4ea3-9c3c-7f0768a257b2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:36:46 crc kubenswrapper[4806]: E1127 10:36:46.959092 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified\\\"\"" pod="openstack/ovn-controller-l82sn" podUID="2cb892e8-358b-477b-8741-4211a599bf6c" Nov 27 10:36:47 crc kubenswrapper[4806]: E1127 10:36:47.457604 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Nov 27 10:36:47 crc kubenswrapper[4806]: E1127 10:36:47.457703 4806 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Nov 27 10:36:47 crc kubenswrapper[4806]: E1127 10:36:47.457884 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kh2rj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(beb345b2-15fd-4db1-8ea0-c862b7820fc0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 27 10:36:47 crc kubenswrapper[4806]: E1127 10:36:47.458992 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="beb345b2-15fd-4db1-8ea0-c862b7820fc0" Nov 27 10:36:47 crc kubenswrapper[4806]: I1127 10:36:47.964913 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" event={"ID":"e5070f2a-0a09-4ff4-b8f6-5978268548d0","Type":"ContainerStarted","Data":"0f1e0d341c4221d6beb8d8aa112bd4aeef092f1f4895b8f576cef478facb28b7"} Nov 27 10:36:47 crc kubenswrapper[4806]: I1127 10:36:47.967894 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:47 crc kubenswrapper[4806]: I1127 10:36:47.965084 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" podUID="e5070f2a-0a09-4ff4-b8f6-5978268548d0" containerName="dnsmasq-dns" containerID="cri-o://0f1e0d341c4221d6beb8d8aa112bd4aeef092f1f4895b8f576cef478facb28b7" gracePeriod=10 Nov 27 10:36:47 crc kubenswrapper[4806]: I1127 10:36:47.971004 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d9a162dd-03ba-4444-b03c-bf34c535713a","Type":"ContainerStarted","Data":"3f434f648342995a8a41507be91f168f10c4d9e875c2b5e1c21d57c10a3c7bfa"} Nov 27 10:36:47 crc kubenswrapper[4806]: I1127 10:36:47.971160 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 27 10:36:47 crc kubenswrapper[4806]: E1127 10:36:47.971922 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="beb345b2-15fd-4db1-8ea0-c862b7820fc0" Nov 27 10:36:47 crc kubenswrapper[4806]: I1127 10:36:47.992283 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" podStartSLOduration=28.49062643 podStartE2EDuration="38.992252466s" podCreationTimestamp="2025-11-27 10:36:09 +0000 UTC" firstStartedPulling="2025-11-27 10:36:16.200462843 +0000 UTC m=+880.787053607" lastFinishedPulling="2025-11-27 10:36:26.702088879 +0000 UTC m=+891.288679643" observedRunningTime="2025-11-27 10:36:47.984299035 +0000 UTC m=+912.570889799" watchObservedRunningTime="2025-11-27 10:36:47.992252466 +0000 UTC m=+912.578843230" Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.046100 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=15.098374443 podStartE2EDuration="34.046039527s" podCreationTimestamp="2025-11-27 10:36:14 +0000 UTC" firstStartedPulling="2025-11-27 10:36:27.1693469 +0000 UTC m=+891.755937664" lastFinishedPulling="2025-11-27 10:36:46.117011944 +0000 UTC m=+910.703602748" observedRunningTime="2025-11-27 10:36:48.036342549 +0000 UTC m=+912.622933303" watchObservedRunningTime="2025-11-27 10:36:48.046039527 +0000 UTC m=+912.632630291" Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.606034 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.729030 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5070f2a-0a09-4ff4-b8f6-5978268548d0-dns-svc\") pod \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\" (UID: \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\") " Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.729088 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5070f2a-0a09-4ff4-b8f6-5978268548d0-config\") pod \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\" (UID: \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\") " Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.729162 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvnfk\" (UniqueName: \"kubernetes.io/projected/e5070f2a-0a09-4ff4-b8f6-5978268548d0-kube-api-access-wvnfk\") pod \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\" (UID: \"e5070f2a-0a09-4ff4-b8f6-5978268548d0\") " Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.733968 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5070f2a-0a09-4ff4-b8f6-5978268548d0-kube-api-access-wvnfk" (OuterVolumeSpecName: "kube-api-access-wvnfk") pod "e5070f2a-0a09-4ff4-b8f6-5978268548d0" (UID: "e5070f2a-0a09-4ff4-b8f6-5978268548d0"). InnerVolumeSpecName "kube-api-access-wvnfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.800969 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5070f2a-0a09-4ff4-b8f6-5978268548d0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e5070f2a-0a09-4ff4-b8f6-5978268548d0" (UID: "e5070f2a-0a09-4ff4-b8f6-5978268548d0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.832357 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5070f2a-0a09-4ff4-b8f6-5978268548d0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.832407 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvnfk\" (UniqueName: \"kubernetes.io/projected/e5070f2a-0a09-4ff4-b8f6-5978268548d0-kube-api-access-wvnfk\") on node \"crc\" DevicePath \"\"" Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.838818 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5070f2a-0a09-4ff4-b8f6-5978268548d0-config" (OuterVolumeSpecName: "config") pod "e5070f2a-0a09-4ff4-b8f6-5978268548d0" (UID: "e5070f2a-0a09-4ff4-b8f6-5978268548d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.934321 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5070f2a-0a09-4ff4-b8f6-5978268548d0-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.984977 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"08137128-e0d1-49d2-abf1-1242c1cb58ce","Type":"ContainerStarted","Data":"e7c3736208836a8b527976f788183343ce7a097c822f342b20cd72859c3b83dd"} Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.987320 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"954f26d7-2bdc-45e3-a07b-d9924469f0a1","Type":"ContainerStarted","Data":"96a9494ee6564f52a58e01b832a4f8f8d3d7c39e07840c0adfcef6d9414c5d1b"} Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.992806 4806 generic.go:334] "Generic (PLEG): container finished" podID="e5070f2a-0a09-4ff4-b8f6-5978268548d0" containerID="0f1e0d341c4221d6beb8d8aa112bd4aeef092f1f4895b8f576cef478facb28b7" exitCode=0 Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.993089 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" event={"ID":"e5070f2a-0a09-4ff4-b8f6-5978268548d0","Type":"ContainerDied","Data":"0f1e0d341c4221d6beb8d8aa112bd4aeef092f1f4895b8f576cef478facb28b7"} Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.993126 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" event={"ID":"e5070f2a-0a09-4ff4-b8f6-5978268548d0","Type":"ContainerDied","Data":"af935e12933b42c76c2fb28ab0529071a24ecb4b936801377f7b242facf7aadf"} Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.993149 4806 scope.go:117] "RemoveContainer" containerID="0f1e0d341c4221d6beb8d8aa112bd4aeef092f1f4895b8f576cef478facb28b7" Nov 27 10:36:48 crc kubenswrapper[4806]: I1127 10:36:48.993129 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-v8mkn" Nov 27 10:36:49 crc kubenswrapper[4806]: I1127 10:36:49.046733 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-v8mkn"] Nov 27 10:36:49 crc kubenswrapper[4806]: I1127 10:36:49.055804 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-v8mkn"] Nov 27 10:36:50 crc kubenswrapper[4806]: I1127 10:36:50.130947 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5070f2a-0a09-4ff4-b8f6-5978268548d0" path="/var/lib/kubelet/pods/e5070f2a-0a09-4ff4-b8f6-5978268548d0/volumes" Nov 27 10:36:50 crc kubenswrapper[4806]: I1127 10:36:50.688156 4806 scope.go:117] "RemoveContainer" containerID="d9077c1972fcc3000136b40d14f8c5cbbd17976f663763ce0eb41fc92791790b" Nov 27 10:36:50 crc kubenswrapper[4806]: I1127 10:36:50.729809 4806 scope.go:117] "RemoveContainer" containerID="0f1e0d341c4221d6beb8d8aa112bd4aeef092f1f4895b8f576cef478facb28b7" Nov 27 10:36:50 crc kubenswrapper[4806]: E1127 10:36:50.730454 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f1e0d341c4221d6beb8d8aa112bd4aeef092f1f4895b8f576cef478facb28b7\": container with ID starting with 0f1e0d341c4221d6beb8d8aa112bd4aeef092f1f4895b8f576cef478facb28b7 not found: ID does not exist" containerID="0f1e0d341c4221d6beb8d8aa112bd4aeef092f1f4895b8f576cef478facb28b7" Nov 27 10:36:50 crc kubenswrapper[4806]: I1127 10:36:50.730516 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1e0d341c4221d6beb8d8aa112bd4aeef092f1f4895b8f576cef478facb28b7"} err="failed to get container status \"0f1e0d341c4221d6beb8d8aa112bd4aeef092f1f4895b8f576cef478facb28b7\": rpc error: code = NotFound desc = could not find container \"0f1e0d341c4221d6beb8d8aa112bd4aeef092f1f4895b8f576cef478facb28b7\": container with ID starting with 0f1e0d341c4221d6beb8d8aa112bd4aeef092f1f4895b8f576cef478facb28b7 not found: ID does not exist" Nov 27 10:36:50 crc kubenswrapper[4806]: I1127 10:36:50.730559 4806 scope.go:117] "RemoveContainer" containerID="d9077c1972fcc3000136b40d14f8c5cbbd17976f663763ce0eb41fc92791790b" Nov 27 10:36:50 crc kubenswrapper[4806]: E1127 10:36:50.731067 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9077c1972fcc3000136b40d14f8c5cbbd17976f663763ce0eb41fc92791790b\": container with ID starting with d9077c1972fcc3000136b40d14f8c5cbbd17976f663763ce0eb41fc92791790b not found: ID does not exist" containerID="d9077c1972fcc3000136b40d14f8c5cbbd17976f663763ce0eb41fc92791790b" Nov 27 10:36:50 crc kubenswrapper[4806]: I1127 10:36:50.731090 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9077c1972fcc3000136b40d14f8c5cbbd17976f663763ce0eb41fc92791790b"} err="failed to get container status \"d9077c1972fcc3000136b40d14f8c5cbbd17976f663763ce0eb41fc92791790b\": rpc error: code = NotFound desc = could not find container \"d9077c1972fcc3000136b40d14f8c5cbbd17976f663763ce0eb41fc92791790b\": container with ID starting with d9077c1972fcc3000136b40d14f8c5cbbd17976f663763ce0eb41fc92791790b not found: ID does not exist" Nov 27 10:36:51 crc kubenswrapper[4806]: E1127 10:36:50.999877 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="08068118-444f-4ea3-9c3c-7f0768a257b2" Nov 27 10:36:51 crc kubenswrapper[4806]: E1127 10:36:51.000039 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="d8134ca5-1cd3-4793-875e-8ea7bde9c407" Nov 27 10:36:51 crc kubenswrapper[4806]: I1127 10:36:51.020465 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d8134ca5-1cd3-4793-875e-8ea7bde9c407","Type":"ContainerStarted","Data":"734c5ea36f1068f3b4a0a5f2fcd5e17ce0ea57ac4556c470df57c68ce871cd6e"} Nov 27 10:36:51 crc kubenswrapper[4806]: E1127 10:36:51.022308 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="d8134ca5-1cd3-4793-875e-8ea7bde9c407" Nov 27 10:36:51 crc kubenswrapper[4806]: I1127 10:36:51.022209 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"08068118-444f-4ea3-9c3c-7f0768a257b2","Type":"ContainerStarted","Data":"bda9e1a4197f5c3edef67431b10e7c5d210eefcebb119311f3cf3ecc2d7df28b"} Nov 27 10:36:51 crc kubenswrapper[4806]: E1127 10:36:51.030440 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="08068118-444f-4ea3-9c3c-7f0768a257b2" Nov 27 10:36:52 crc kubenswrapper[4806]: E1127 10:36:52.033549 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="08068118-444f-4ea3-9c3c-7f0768a257b2" Nov 27 10:36:52 crc kubenswrapper[4806]: E1127 10:36:52.033995 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-sb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="d8134ca5-1cd3-4793-875e-8ea7bde9c407" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.181346 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lw6vw"] Nov 27 10:36:52 crc kubenswrapper[4806]: E1127 10:36:52.181968 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5070f2a-0a09-4ff4-b8f6-5978268548d0" containerName="dnsmasq-dns" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.181988 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5070f2a-0a09-4ff4-b8f6-5978268548d0" containerName="dnsmasq-dns" Nov 27 10:36:52 crc kubenswrapper[4806]: E1127 10:36:52.182016 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5070f2a-0a09-4ff4-b8f6-5978268548d0" containerName="init" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.182023 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5070f2a-0a09-4ff4-b8f6-5978268548d0" containerName="init" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.182195 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5070f2a-0a09-4ff4-b8f6-5978268548d0" containerName="dnsmasq-dns" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.183650 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.195473 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lw6vw"] Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.200328 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2be71d68-ee0c-4f3b-b823-dbf4df159889-utilities\") pod \"redhat-marketplace-lw6vw\" (UID: \"2be71d68-ee0c-4f3b-b823-dbf4df159889\") " pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.200850 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2be71d68-ee0c-4f3b-b823-dbf4df159889-catalog-content\") pod \"redhat-marketplace-lw6vw\" (UID: \"2be71d68-ee0c-4f3b-b823-dbf4df159889\") " pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.200954 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhc9m\" (UniqueName: \"kubernetes.io/projected/2be71d68-ee0c-4f3b-b823-dbf4df159889-kube-api-access-lhc9m\") pod \"redhat-marketplace-lw6vw\" (UID: \"2be71d68-ee0c-4f3b-b823-dbf4df159889\") " pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.302950 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2be71d68-ee0c-4f3b-b823-dbf4df159889-catalog-content\") pod \"redhat-marketplace-lw6vw\" (UID: \"2be71d68-ee0c-4f3b-b823-dbf4df159889\") " pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.303043 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhc9m\" (UniqueName: \"kubernetes.io/projected/2be71d68-ee0c-4f3b-b823-dbf4df159889-kube-api-access-lhc9m\") pod \"redhat-marketplace-lw6vw\" (UID: \"2be71d68-ee0c-4f3b-b823-dbf4df159889\") " pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.303102 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2be71d68-ee0c-4f3b-b823-dbf4df159889-utilities\") pod \"redhat-marketplace-lw6vw\" (UID: \"2be71d68-ee0c-4f3b-b823-dbf4df159889\") " pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.303756 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2be71d68-ee0c-4f3b-b823-dbf4df159889-utilities\") pod \"redhat-marketplace-lw6vw\" (UID: \"2be71d68-ee0c-4f3b-b823-dbf4df159889\") " pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.303761 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2be71d68-ee0c-4f3b-b823-dbf4df159889-catalog-content\") pod \"redhat-marketplace-lw6vw\" (UID: \"2be71d68-ee0c-4f3b-b823-dbf4df159889\") " pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.322651 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhc9m\" (UniqueName: \"kubernetes.io/projected/2be71d68-ee0c-4f3b-b823-dbf4df159889-kube-api-access-lhc9m\") pod \"redhat-marketplace-lw6vw\" (UID: \"2be71d68-ee0c-4f3b-b823-dbf4df159889\") " pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.501209 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:36:52 crc kubenswrapper[4806]: I1127 10:36:52.950429 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lw6vw"] Nov 27 10:36:52 crc kubenswrapper[4806]: W1127 10:36:52.954027 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2be71d68_ee0c_4f3b_b823_dbf4df159889.slice/crio-4878f714f1d5c5ea0bd42010f4adf07e65bbcd0f842f335ebf354dfa31c6b5ee WatchSource:0}: Error finding container 4878f714f1d5c5ea0bd42010f4adf07e65bbcd0f842f335ebf354dfa31c6b5ee: Status 404 returned error can't find the container with id 4878f714f1d5c5ea0bd42010f4adf07e65bbcd0f842f335ebf354dfa31c6b5ee Nov 27 10:36:53 crc kubenswrapper[4806]: I1127 10:36:53.039121 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lw6vw" event={"ID":"2be71d68-ee0c-4f3b-b823-dbf4df159889","Type":"ContainerStarted","Data":"4878f714f1d5c5ea0bd42010f4adf07e65bbcd0f842f335ebf354dfa31c6b5ee"} Nov 27 10:36:54 crc kubenswrapper[4806]: I1127 10:36:54.048560 4806 generic.go:334] "Generic (PLEG): container finished" podID="2be71d68-ee0c-4f3b-b823-dbf4df159889" containerID="e68987f674ae79f9e9719f9a63a5628fe96df5449d0d3a9504bda2b85093a6bd" exitCode=0 Nov 27 10:36:54 crc kubenswrapper[4806]: I1127 10:36:54.048771 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lw6vw" event={"ID":"2be71d68-ee0c-4f3b-b823-dbf4df159889","Type":"ContainerDied","Data":"e68987f674ae79f9e9719f9a63a5628fe96df5449d0d3a9504bda2b85093a6bd"} Nov 27 10:36:54 crc kubenswrapper[4806]: I1127 10:36:54.711207 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 27 10:36:55 crc kubenswrapper[4806]: I1127 10:36:55.062123 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lw6vw" event={"ID":"2be71d68-ee0c-4f3b-b823-dbf4df159889","Type":"ContainerStarted","Data":"9ee6be6198d663f7ae8c1d7530ef983c4a6d1afeccc29cb2e88b24582d2a5836"} Nov 27 10:36:56 crc kubenswrapper[4806]: I1127 10:36:56.071977 4806 generic.go:334] "Generic (PLEG): container finished" podID="2be71d68-ee0c-4f3b-b823-dbf4df159889" containerID="9ee6be6198d663f7ae8c1d7530ef983c4a6d1afeccc29cb2e88b24582d2a5836" exitCode=0 Nov 27 10:36:56 crc kubenswrapper[4806]: I1127 10:36:56.072259 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lw6vw" event={"ID":"2be71d68-ee0c-4f3b-b823-dbf4df159889","Type":"ContainerDied","Data":"9ee6be6198d663f7ae8c1d7530ef983c4a6d1afeccc29cb2e88b24582d2a5836"} Nov 27 10:36:59 crc kubenswrapper[4806]: I1127 10:36:59.118183 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lw6vw" event={"ID":"2be71d68-ee0c-4f3b-b823-dbf4df159889","Type":"ContainerStarted","Data":"87500a5ee46877d9c49b1d06b9badeb961e01b13f03bd27e1a8b0da86ccf2a57"} Nov 27 10:36:59 crc kubenswrapper[4806]: I1127 10:36:59.171363 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lw6vw" podStartSLOduration=2.489342288 podStartE2EDuration="7.171342185s" podCreationTimestamp="2025-11-27 10:36:52 +0000 UTC" firstStartedPulling="2025-11-27 10:36:54.051119705 +0000 UTC m=+918.637710469" lastFinishedPulling="2025-11-27 10:36:58.733119602 +0000 UTC m=+923.319710366" observedRunningTime="2025-11-27 10:36:59.165420581 +0000 UTC m=+923.752011345" watchObservedRunningTime="2025-11-27 10:36:59.171342185 +0000 UTC m=+923.757932939" Nov 27 10:37:00 crc kubenswrapper[4806]: I1127 10:37:00.135813 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l82sn" event={"ID":"2cb892e8-358b-477b-8741-4211a599bf6c","Type":"ContainerStarted","Data":"847d52deff4922fd3ea9d9befd9903a62dde151522681807584010ee5835be4b"} Nov 27 10:37:00 crc kubenswrapper[4806]: I1127 10:37:00.137686 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-l82sn" Nov 27 10:37:00 crc kubenswrapper[4806]: I1127 10:37:00.144178 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f769m" event={"ID":"d814f280-e3ac-426c-bafa-4a4ed3f51320","Type":"ContainerStarted","Data":"e23aa7d257ddad2da3a2248c19933ef0c9e082daf8bdcca03c8bd5be204e2a72"} Nov 27 10:37:00 crc kubenswrapper[4806]: I1127 10:37:00.146854 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b30ea2f1-af76-4a43-842c-577ff4f35a20","Type":"ContainerStarted","Data":"4071b21ddadcba1db30be12fa0a473aa0c38ed1ea363fd483867b790b53771ce"} Nov 27 10:37:00 crc kubenswrapper[4806]: I1127 10:37:00.150091 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"46cb1b0e-a6ab-467a-9f49-f8938c164253","Type":"ContainerStarted","Data":"1077cee8ffc39ffd6e1946d41bf42a9af29d583db3acbcff78df947a556da263"} Nov 27 10:37:00 crc kubenswrapper[4806]: I1127 10:37:00.152683 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"beb345b2-15fd-4db1-8ea0-c862b7820fc0","Type":"ContainerStarted","Data":"1356dc8aaca88cf9d159fd1ff7cd38621ef0d9cca9817734724a264b1475f593"} Nov 27 10:37:00 crc kubenswrapper[4806]: I1127 10:37:00.153213 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 27 10:37:00 crc kubenswrapper[4806]: I1127 10:37:00.194438 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-l82sn" podStartSLOduration=9.84998818 podStartE2EDuration="41.194400726s" podCreationTimestamp="2025-11-27 10:36:19 +0000 UTC" firstStartedPulling="2025-11-27 10:36:27.641782235 +0000 UTC m=+892.228372999" lastFinishedPulling="2025-11-27 10:36:58.986194781 +0000 UTC m=+923.572785545" observedRunningTime="2025-11-27 10:37:00.168030415 +0000 UTC m=+924.754621169" watchObservedRunningTime="2025-11-27 10:37:00.194400726 +0000 UTC m=+924.780991480" Nov 27 10:37:00 crc kubenswrapper[4806]: I1127 10:37:00.290598 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=12.016158031 podStartE2EDuration="44.290567172s" podCreationTimestamp="2025-11-27 10:36:16 +0000 UTC" firstStartedPulling="2025-11-27 10:36:27.225013152 +0000 UTC m=+891.811603916" lastFinishedPulling="2025-11-27 10:36:59.499422293 +0000 UTC m=+924.086013057" observedRunningTime="2025-11-27 10:37:00.282428146 +0000 UTC m=+924.869018930" watchObservedRunningTime="2025-11-27 10:37:00.290567172 +0000 UTC m=+924.877157946" Nov 27 10:37:01 crc kubenswrapper[4806]: I1127 10:37:01.161810 4806 generic.go:334] "Generic (PLEG): container finished" podID="d814f280-e3ac-426c-bafa-4a4ed3f51320" containerID="e23aa7d257ddad2da3a2248c19933ef0c9e082daf8bdcca03c8bd5be204e2a72" exitCode=0 Nov 27 10:37:01 crc kubenswrapper[4806]: I1127 10:37:01.163235 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f769m" event={"ID":"d814f280-e3ac-426c-bafa-4a4ed3f51320","Type":"ContainerDied","Data":"e23aa7d257ddad2da3a2248c19933ef0c9e082daf8bdcca03c8bd5be204e2a72"} Nov 27 10:37:02 crc kubenswrapper[4806]: I1127 10:37:02.173113 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f769m" event={"ID":"d814f280-e3ac-426c-bafa-4a4ed3f51320","Type":"ContainerStarted","Data":"218d01130896b9c60ee5f2b1d87c0e63fafaa2194e7e5105efa3f247f4d67ed3"} Nov 27 10:37:02 crc kubenswrapper[4806]: I1127 10:37:02.173542 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-f769m" event={"ID":"d814f280-e3ac-426c-bafa-4a4ed3f51320","Type":"ContainerStarted","Data":"98a042463be38b65962c969f1af38c00bfaa1629df17d397a96ddd6d39e42b2f"} Nov 27 10:37:02 crc kubenswrapper[4806]: I1127 10:37:02.173819 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:37:02 crc kubenswrapper[4806]: I1127 10:37:02.502006 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:37:02 crc kubenswrapper[4806]: I1127 10:37:02.502242 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:37:02 crc kubenswrapper[4806]: I1127 10:37:02.575542 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:37:02 crc kubenswrapper[4806]: I1127 10:37:02.606873 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-f769m" podStartSLOduration=15.820836542 podStartE2EDuration="43.606849295s" podCreationTimestamp="2025-11-27 10:36:19 +0000 UTC" firstStartedPulling="2025-11-27 10:36:31.203271364 +0000 UTC m=+895.789862128" lastFinishedPulling="2025-11-27 10:36:58.989284117 +0000 UTC m=+923.575874881" observedRunningTime="2025-11-27 10:37:02.204837887 +0000 UTC m=+926.791428641" watchObservedRunningTime="2025-11-27 10:37:02.606849295 +0000 UTC m=+927.193440059" Nov 27 10:37:03 crc kubenswrapper[4806]: I1127 10:37:03.183560 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:37:03 crc kubenswrapper[4806]: I1127 10:37:03.233746 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:37:03 crc kubenswrapper[4806]: I1127 10:37:03.294664 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lw6vw"] Nov 27 10:37:04 crc kubenswrapper[4806]: I1127 10:37:04.192776 4806 generic.go:334] "Generic (PLEG): container finished" podID="b30ea2f1-af76-4a43-842c-577ff4f35a20" containerID="4071b21ddadcba1db30be12fa0a473aa0c38ed1ea363fd483867b790b53771ce" exitCode=0 Nov 27 10:37:04 crc kubenswrapper[4806]: I1127 10:37:04.192869 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b30ea2f1-af76-4a43-842c-577ff4f35a20","Type":"ContainerDied","Data":"4071b21ddadcba1db30be12fa0a473aa0c38ed1ea363fd483867b790b53771ce"} Nov 27 10:37:04 crc kubenswrapper[4806]: I1127 10:37:04.196535 4806 generic.go:334] "Generic (PLEG): container finished" podID="46cb1b0e-a6ab-467a-9f49-f8938c164253" containerID="1077cee8ffc39ffd6e1946d41bf42a9af29d583db3acbcff78df947a556da263" exitCode=0 Nov 27 10:37:04 crc kubenswrapper[4806]: I1127 10:37:04.196709 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"46cb1b0e-a6ab-467a-9f49-f8938c164253","Type":"ContainerDied","Data":"1077cee8ffc39ffd6e1946d41bf42a9af29d583db3acbcff78df947a556da263"} Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.208335 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b30ea2f1-af76-4a43-842c-577ff4f35a20","Type":"ContainerStarted","Data":"510c1282f3eb9a9fd4eaec9d4c26cc94a30c852bf8a89b8473e19c40c795263f"} Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.211156 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"46cb1b0e-a6ab-467a-9f49-f8938c164253","Type":"ContainerStarted","Data":"092bace9c5ad99ac712fee7ee8315338efd2f4230aa1fd89c750f6cdcc480fb6"} Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.211309 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lw6vw" podUID="2be71d68-ee0c-4f3b-b823-dbf4df159889" containerName="registry-server" containerID="cri-o://87500a5ee46877d9c49b1d06b9badeb961e01b13f03bd27e1a8b0da86ccf2a57" gracePeriod=2 Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.268613 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=21.78519959 podStartE2EDuration="54.268583708s" podCreationTimestamp="2025-11-27 10:36:11 +0000 UTC" firstStartedPulling="2025-11-27 10:36:26.503543133 +0000 UTC m=+891.090133907" lastFinishedPulling="2025-11-27 10:36:58.986927261 +0000 UTC m=+923.573518025" observedRunningTime="2025-11-27 10:37:05.238662299 +0000 UTC m=+929.825253073" watchObservedRunningTime="2025-11-27 10:37:05.268583708 +0000 UTC m=+929.855174472" Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.275418 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=20.644006642 podStartE2EDuration="53.275395728s" podCreationTimestamp="2025-11-27 10:36:12 +0000 UTC" firstStartedPulling="2025-11-27 10:36:26.480559392 +0000 UTC m=+891.067150156" lastFinishedPulling="2025-11-27 10:36:59.111948478 +0000 UTC m=+923.698539242" observedRunningTime="2025-11-27 10:37:05.268421884 +0000 UTC m=+929.855012658" watchObservedRunningTime="2025-11-27 10:37:05.275395728 +0000 UTC m=+929.861986492" Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.594133 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.759705 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2be71d68-ee0c-4f3b-b823-dbf4df159889-catalog-content\") pod \"2be71d68-ee0c-4f3b-b823-dbf4df159889\" (UID: \"2be71d68-ee0c-4f3b-b823-dbf4df159889\") " Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.759863 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2be71d68-ee0c-4f3b-b823-dbf4df159889-utilities\") pod \"2be71d68-ee0c-4f3b-b823-dbf4df159889\" (UID: \"2be71d68-ee0c-4f3b-b823-dbf4df159889\") " Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.759974 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhc9m\" (UniqueName: \"kubernetes.io/projected/2be71d68-ee0c-4f3b-b823-dbf4df159889-kube-api-access-lhc9m\") pod \"2be71d68-ee0c-4f3b-b823-dbf4df159889\" (UID: \"2be71d68-ee0c-4f3b-b823-dbf4df159889\") " Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.760850 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2be71d68-ee0c-4f3b-b823-dbf4df159889-utilities" (OuterVolumeSpecName: "utilities") pod "2be71d68-ee0c-4f3b-b823-dbf4df159889" (UID: "2be71d68-ee0c-4f3b-b823-dbf4df159889"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.761610 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2be71d68-ee0c-4f3b-b823-dbf4df159889-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.765448 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2be71d68-ee0c-4f3b-b823-dbf4df159889-kube-api-access-lhc9m" (OuterVolumeSpecName: "kube-api-access-lhc9m") pod "2be71d68-ee0c-4f3b-b823-dbf4df159889" (UID: "2be71d68-ee0c-4f3b-b823-dbf4df159889"). InnerVolumeSpecName "kube-api-access-lhc9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.778714 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2be71d68-ee0c-4f3b-b823-dbf4df159889-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2be71d68-ee0c-4f3b-b823-dbf4df159889" (UID: "2be71d68-ee0c-4f3b-b823-dbf4df159889"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.863032 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2be71d68-ee0c-4f3b-b823-dbf4df159889-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:05 crc kubenswrapper[4806]: I1127 10:37:05.863075 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhc9m\" (UniqueName: \"kubernetes.io/projected/2be71d68-ee0c-4f3b-b823-dbf4df159889-kube-api-access-lhc9m\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.220085 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d8134ca5-1cd3-4793-875e-8ea7bde9c407","Type":"ContainerStarted","Data":"f50642913a8bfa7e7a26ee28fd3b8c97438bc275adec43b9e9527e0887080104"} Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.222618 4806 generic.go:334] "Generic (PLEG): container finished" podID="2be71d68-ee0c-4f3b-b823-dbf4df159889" containerID="87500a5ee46877d9c49b1d06b9badeb961e01b13f03bd27e1a8b0da86ccf2a57" exitCode=0 Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.222711 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lw6vw" Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.222701 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lw6vw" event={"ID":"2be71d68-ee0c-4f3b-b823-dbf4df159889","Type":"ContainerDied","Data":"87500a5ee46877d9c49b1d06b9badeb961e01b13f03bd27e1a8b0da86ccf2a57"} Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.222830 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lw6vw" event={"ID":"2be71d68-ee0c-4f3b-b823-dbf4df159889","Type":"ContainerDied","Data":"4878f714f1d5c5ea0bd42010f4adf07e65bbcd0f842f335ebf354dfa31c6b5ee"} Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.222853 4806 scope.go:117] "RemoveContainer" containerID="87500a5ee46877d9c49b1d06b9badeb961e01b13f03bd27e1a8b0da86ccf2a57" Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.243547 4806 scope.go:117] "RemoveContainer" containerID="9ee6be6198d663f7ae8c1d7530ef983c4a6d1afeccc29cb2e88b24582d2a5836" Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.260817 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=7.505127166 podStartE2EDuration="44.260793023s" podCreationTimestamp="2025-11-27 10:36:22 +0000 UTC" firstStartedPulling="2025-11-27 10:36:28.27487363 +0000 UTC m=+892.861464394" lastFinishedPulling="2025-11-27 10:37:05.030539477 +0000 UTC m=+929.617130251" observedRunningTime="2025-11-27 10:37:06.25456229 +0000 UTC m=+930.841153074" watchObservedRunningTime="2025-11-27 10:37:06.260793023 +0000 UTC m=+930.847383787" Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.277596 4806 scope.go:117] "RemoveContainer" containerID="e68987f674ae79f9e9719f9a63a5628fe96df5449d0d3a9504bda2b85093a6bd" Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.304882 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lw6vw"] Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.320604 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lw6vw"] Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.324535 4806 scope.go:117] "RemoveContainer" containerID="87500a5ee46877d9c49b1d06b9badeb961e01b13f03bd27e1a8b0da86ccf2a57" Nov 27 10:37:06 crc kubenswrapper[4806]: E1127 10:37:06.327421 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87500a5ee46877d9c49b1d06b9badeb961e01b13f03bd27e1a8b0da86ccf2a57\": container with ID starting with 87500a5ee46877d9c49b1d06b9badeb961e01b13f03bd27e1a8b0da86ccf2a57 not found: ID does not exist" containerID="87500a5ee46877d9c49b1d06b9badeb961e01b13f03bd27e1a8b0da86ccf2a57" Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.327491 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87500a5ee46877d9c49b1d06b9badeb961e01b13f03bd27e1a8b0da86ccf2a57"} err="failed to get container status \"87500a5ee46877d9c49b1d06b9badeb961e01b13f03bd27e1a8b0da86ccf2a57\": rpc error: code = NotFound desc = could not find container \"87500a5ee46877d9c49b1d06b9badeb961e01b13f03bd27e1a8b0da86ccf2a57\": container with ID starting with 87500a5ee46877d9c49b1d06b9badeb961e01b13f03bd27e1a8b0da86ccf2a57 not found: ID does not exist" Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.327527 4806 scope.go:117] "RemoveContainer" containerID="9ee6be6198d663f7ae8c1d7530ef983c4a6d1afeccc29cb2e88b24582d2a5836" Nov 27 10:37:06 crc kubenswrapper[4806]: E1127 10:37:06.329494 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ee6be6198d663f7ae8c1d7530ef983c4a6d1afeccc29cb2e88b24582d2a5836\": container with ID starting with 9ee6be6198d663f7ae8c1d7530ef983c4a6d1afeccc29cb2e88b24582d2a5836 not found: ID does not exist" containerID="9ee6be6198d663f7ae8c1d7530ef983c4a6d1afeccc29cb2e88b24582d2a5836" Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.329540 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ee6be6198d663f7ae8c1d7530ef983c4a6d1afeccc29cb2e88b24582d2a5836"} err="failed to get container status \"9ee6be6198d663f7ae8c1d7530ef983c4a6d1afeccc29cb2e88b24582d2a5836\": rpc error: code = NotFound desc = could not find container \"9ee6be6198d663f7ae8c1d7530ef983c4a6d1afeccc29cb2e88b24582d2a5836\": container with ID starting with 9ee6be6198d663f7ae8c1d7530ef983c4a6d1afeccc29cb2e88b24582d2a5836 not found: ID does not exist" Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.329573 4806 scope.go:117] "RemoveContainer" containerID="e68987f674ae79f9e9719f9a63a5628fe96df5449d0d3a9504bda2b85093a6bd" Nov 27 10:37:06 crc kubenswrapper[4806]: E1127 10:37:06.329893 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e68987f674ae79f9e9719f9a63a5628fe96df5449d0d3a9504bda2b85093a6bd\": container with ID starting with e68987f674ae79f9e9719f9a63a5628fe96df5449d0d3a9504bda2b85093a6bd not found: ID does not exist" containerID="e68987f674ae79f9e9719f9a63a5628fe96df5449d0d3a9504bda2b85093a6bd" Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.329922 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e68987f674ae79f9e9719f9a63a5628fe96df5449d0d3a9504bda2b85093a6bd"} err="failed to get container status \"e68987f674ae79f9e9719f9a63a5628fe96df5449d0d3a9504bda2b85093a6bd\": rpc error: code = NotFound desc = could not find container \"e68987f674ae79f9e9719f9a63a5628fe96df5449d0d3a9504bda2b85093a6bd\": container with ID starting with e68987f674ae79f9e9719f9a63a5628fe96df5449d0d3a9504bda2b85093a6bd not found: ID does not exist" Nov 27 10:37:06 crc kubenswrapper[4806]: I1127 10:37:06.456565 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.128490 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2be71d68-ee0c-4f3b-b823-dbf4df159889" path="/var/lib/kubelet/pods/2be71d68-ee0c-4f3b-b823-dbf4df159889/volumes" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.251017 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"08068118-444f-4ea3-9c3c-7f0768a257b2","Type":"ContainerStarted","Data":"26ab46753260ced1efe57b1b024aa3045c90f8421b51aba6222681f253da010b"} Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.253826 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pc2bp"] Nov 27 10:37:08 crc kubenswrapper[4806]: E1127 10:37:08.254282 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2be71d68-ee0c-4f3b-b823-dbf4df159889" containerName="extract-content" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.254304 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2be71d68-ee0c-4f3b-b823-dbf4df159889" containerName="extract-content" Nov 27 10:37:08 crc kubenswrapper[4806]: E1127 10:37:08.254328 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2be71d68-ee0c-4f3b-b823-dbf4df159889" containerName="extract-utilities" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.254336 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2be71d68-ee0c-4f3b-b823-dbf4df159889" containerName="extract-utilities" Nov 27 10:37:08 crc kubenswrapper[4806]: E1127 10:37:08.254365 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2be71d68-ee0c-4f3b-b823-dbf4df159889" containerName="registry-server" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.254373 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2be71d68-ee0c-4f3b-b823-dbf4df159889" containerName="registry-server" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.257330 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="2be71d68-ee0c-4f3b-b823-dbf4df159889" containerName="registry-server" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.261116 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.278754 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pc2bp"] Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.284225 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=9.112852665 podStartE2EDuration="47.284194955s" podCreationTimestamp="2025-11-27 10:36:21 +0000 UTC" firstStartedPulling="2025-11-27 10:36:29.401426586 +0000 UTC m=+893.988017350" lastFinishedPulling="2025-11-27 10:37:07.572768876 +0000 UTC m=+932.159359640" observedRunningTime="2025-11-27 10:37:08.278881047 +0000 UTC m=+932.865471801" watchObservedRunningTime="2025-11-27 10:37:08.284194955 +0000 UTC m=+932.870785719" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.422962 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a19e7111-3ef1-45a9-ae31-0d53a2a6615f-utilities\") pod \"community-operators-pc2bp\" (UID: \"a19e7111-3ef1-45a9-ae31-0d53a2a6615f\") " pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.423494 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt62j\" (UniqueName: \"kubernetes.io/projected/a19e7111-3ef1-45a9-ae31-0d53a2a6615f-kube-api-access-qt62j\") pod \"community-operators-pc2bp\" (UID: \"a19e7111-3ef1-45a9-ae31-0d53a2a6615f\") " pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.423635 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a19e7111-3ef1-45a9-ae31-0d53a2a6615f-catalog-content\") pod \"community-operators-pc2bp\" (UID: \"a19e7111-3ef1-45a9-ae31-0d53a2a6615f\") " pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.525558 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt62j\" (UniqueName: \"kubernetes.io/projected/a19e7111-3ef1-45a9-ae31-0d53a2a6615f-kube-api-access-qt62j\") pod \"community-operators-pc2bp\" (UID: \"a19e7111-3ef1-45a9-ae31-0d53a2a6615f\") " pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.525920 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a19e7111-3ef1-45a9-ae31-0d53a2a6615f-catalog-content\") pod \"community-operators-pc2bp\" (UID: \"a19e7111-3ef1-45a9-ae31-0d53a2a6615f\") " pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.526037 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a19e7111-3ef1-45a9-ae31-0d53a2a6615f-utilities\") pod \"community-operators-pc2bp\" (UID: \"a19e7111-3ef1-45a9-ae31-0d53a2a6615f\") " pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.526553 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a19e7111-3ef1-45a9-ae31-0d53a2a6615f-catalog-content\") pod \"community-operators-pc2bp\" (UID: \"a19e7111-3ef1-45a9-ae31-0d53a2a6615f\") " pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.526618 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a19e7111-3ef1-45a9-ae31-0d53a2a6615f-utilities\") pod \"community-operators-pc2bp\" (UID: \"a19e7111-3ef1-45a9-ae31-0d53a2a6615f\") " pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.555063 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt62j\" (UniqueName: \"kubernetes.io/projected/a19e7111-3ef1-45a9-ae31-0d53a2a6615f-kube-api-access-qt62j\") pod \"community-operators-pc2bp\" (UID: \"a19e7111-3ef1-45a9-ae31-0d53a2a6615f\") " pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.583424 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.830575 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.832189 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 27 10:37:08 crc kubenswrapper[4806]: I1127 10:37:08.931476 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 27 10:37:09 crc kubenswrapper[4806]: I1127 10:37:09.262502 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pc2bp"] Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.267863 4806 generic.go:334] "Generic (PLEG): container finished" podID="a19e7111-3ef1-45a9-ae31-0d53a2a6615f" containerID="80f923b2e3c21ae9737e9083655af56fa516817555de673eb23f486f6de81a4a" exitCode=0 Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.271015 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pc2bp" event={"ID":"a19e7111-3ef1-45a9-ae31-0d53a2a6615f","Type":"ContainerDied","Data":"80f923b2e3c21ae9737e9083655af56fa516817555de673eb23f486f6de81a4a"} Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.271341 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pc2bp" event={"ID":"a19e7111-3ef1-45a9-ae31-0d53a2a6615f","Type":"ContainerStarted","Data":"75b50fc77d2e8421376171641944f4df790108ed8ad0630fd7646d8e6e8cc153"} Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.483220 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.802613 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-672fg"] Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.804120 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.810843 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.821186 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-672fg"] Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.933842 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-xg9f9"] Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.934969 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.942929 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.949810 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-xg9f9"] Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.970478 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-672fg\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.970544 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpfk5\" (UniqueName: \"kubernetes.io/projected/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-kube-api-access-cpfk5\") pod \"dnsmasq-dns-6bc7876d45-672fg\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.970589 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-672fg\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:10 crc kubenswrapper[4806]: I1127 10:37:10.970647 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-config\") pod \"dnsmasq-dns-6bc7876d45-672fg\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.070460 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.071779 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a81c0343-88d6-42cf-bc8f-5a3b8722a137-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.071846 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a81c0343-88d6-42cf-bc8f-5a3b8722a137-ovn-rundir\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.071885 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-config\") pod \"dnsmasq-dns-6bc7876d45-672fg\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.071923 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a81c0343-88d6-42cf-bc8f-5a3b8722a137-combined-ca-bundle\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.071942 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a81c0343-88d6-42cf-bc8f-5a3b8722a137-ovs-rundir\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.071964 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-672fg\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.072002 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpfk5\" (UniqueName: \"kubernetes.io/projected/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-kube-api-access-cpfk5\") pod \"dnsmasq-dns-6bc7876d45-672fg\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.072034 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77thc\" (UniqueName: \"kubernetes.io/projected/a81c0343-88d6-42cf-bc8f-5a3b8722a137-kube-api-access-77thc\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.072056 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a81c0343-88d6-42cf-bc8f-5a3b8722a137-config\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.072079 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-672fg\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.073232 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-672fg\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.074066 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-672fg\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.074275 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-config\") pod \"dnsmasq-dns-6bc7876d45-672fg\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.111870 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpfk5\" (UniqueName: \"kubernetes.io/projected/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-kube-api-access-cpfk5\") pod \"dnsmasq-dns-6bc7876d45-672fg\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.135673 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.148587 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.173397 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a81c0343-88d6-42cf-bc8f-5a3b8722a137-combined-ca-bundle\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.173446 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a81c0343-88d6-42cf-bc8f-5a3b8722a137-ovs-rundir\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.173553 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77thc\" (UniqueName: \"kubernetes.io/projected/a81c0343-88d6-42cf-bc8f-5a3b8722a137-kube-api-access-77thc\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.173576 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a81c0343-88d6-42cf-bc8f-5a3b8722a137-config\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.173619 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a81c0343-88d6-42cf-bc8f-5a3b8722a137-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.173651 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a81c0343-88d6-42cf-bc8f-5a3b8722a137-ovn-rundir\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.175297 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a81c0343-88d6-42cf-bc8f-5a3b8722a137-ovs-rundir\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.182610 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a81c0343-88d6-42cf-bc8f-5a3b8722a137-config\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.183044 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a81c0343-88d6-42cf-bc8f-5a3b8722a137-ovn-rundir\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.183792 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a81c0343-88d6-42cf-bc8f-5a3b8722a137-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.191919 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a81c0343-88d6-42cf-bc8f-5a3b8722a137-combined-ca-bundle\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.225875 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77thc\" (UniqueName: \"kubernetes.io/projected/a81c0343-88d6-42cf-bc8f-5a3b8722a137-kube-api-access-77thc\") pod \"ovn-controller-metrics-xg9f9\" (UID: \"a81c0343-88d6-42cf-bc8f-5a3b8722a137\") " pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.265500 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-xg9f9" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.282985 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.294949 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-672fg"] Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.346447 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-9hn6t"] Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.348964 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.359496 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.370213 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-9hn6t"] Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.493050 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-dns-svc\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.493385 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.493435 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-config\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.493495 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.493542 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrncd\" (UniqueName: \"kubernetes.io/projected/9aec6519-44e3-499c-bdce-151bf8a13a25-kube-api-access-qrncd\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.595079 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-config\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.595177 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.595232 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrncd\" (UniqueName: \"kubernetes.io/projected/9aec6519-44e3-499c-bdce-151bf8a13a25-kube-api-access-qrncd\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.595286 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-dns-svc\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.595317 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.598448 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-dns-svc\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.598574 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.598833 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.602206 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-config\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.623730 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrncd\" (UniqueName: \"kubernetes.io/projected/9aec6519-44e3-499c-bdce-151bf8a13a25-kube-api-access-qrncd\") pod \"dnsmasq-dns-8554648995-9hn6t\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.684944 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.800848 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-672fg"] Nov 27 10:37:11 crc kubenswrapper[4806]: W1127 10:37:11.824612 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f57f2e3_1d7a_4e6c_a67d_b8c81823310b.slice/crio-fe3f0e2635539cb8cf82dc1a2a76b8aa4750f66265359e769b6fa0a81894b5dd WatchSource:0}: Error finding container fe3f0e2635539cb8cf82dc1a2a76b8aa4750f66265359e769b6fa0a81894b5dd: Status 404 returned error can't find the container with id fe3f0e2635539cb8cf82dc1a2a76b8aa4750f66265359e769b6fa0a81894b5dd Nov 27 10:37:11 crc kubenswrapper[4806]: I1127 10:37:11.940851 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-xg9f9"] Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.173512 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-9hn6t"] Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.289802 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-9hn6t" event={"ID":"9aec6519-44e3-499c-bdce-151bf8a13a25","Type":"ContainerStarted","Data":"fe4cf7cb01cdfec80a130a332a5b62fe1d63bcbfaebe70bdfb79fa5bc015f72e"} Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.292084 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-xg9f9" event={"ID":"a81c0343-88d6-42cf-bc8f-5a3b8722a137","Type":"ContainerStarted","Data":"47049939cc3d428614b9bb56ec5dd53cd00fdfbe0c3123d7537c4a9b3e4c5298"} Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.297627 4806 generic.go:334] "Generic (PLEG): container finished" podID="6f57f2e3-1d7a-4e6c-a67d-b8c81823310b" containerID="ff59c59ea0bf11c03fd13fc074f34fca6311a75a31cb0c7735cb9bff2bb4f834" exitCode=0 Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.297779 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-672fg" event={"ID":"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b","Type":"ContainerDied","Data":"ff59c59ea0bf11c03fd13fc074f34fca6311a75a31cb0c7735cb9bff2bb4f834"} Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.298126 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-672fg" event={"ID":"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b","Type":"ContainerStarted","Data":"fe3f0e2635539cb8cf82dc1a2a76b8aa4750f66265359e769b6fa0a81894b5dd"} Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.392324 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.654204 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.715755 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 27 10:37:12 crc kubenswrapper[4806]: E1127 10:37:12.716086 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f57f2e3-1d7a-4e6c-a67d-b8c81823310b" containerName="init" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.716098 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f57f2e3-1d7a-4e6c-a67d-b8c81823310b" containerName="init" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.716284 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f57f2e3-1d7a-4e6c-a67d-b8c81823310b" containerName="init" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.717140 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.718989 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-dns-svc\") pod \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.719055 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpfk5\" (UniqueName: \"kubernetes.io/projected/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-kube-api-access-cpfk5\") pod \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.719166 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-ovsdbserver-sb\") pod \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.719225 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-config\") pod \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\" (UID: \"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b\") " Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.724180 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.724497 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-9t7np" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.725858 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.733822 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.757585 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-kube-api-access-cpfk5" (OuterVolumeSpecName: "kube-api-access-cpfk5") pod "6f57f2e3-1d7a-4e6c-a67d-b8c81823310b" (UID: "6f57f2e3-1d7a-4e6c-a67d-b8c81823310b"). InnerVolumeSpecName "kube-api-access-cpfk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.801405 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.805070 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6f57f2e3-1d7a-4e6c-a67d-b8c81823310b" (UID: "6f57f2e3-1d7a-4e6c-a67d-b8c81823310b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.826124 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a10ac5e-a435-45ef-b143-f6578010557f-scripts\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.826181 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a10ac5e-a435-45ef-b143-f6578010557f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.826249 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a10ac5e-a435-45ef-b143-f6578010557f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.826274 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a10ac5e-a435-45ef-b143-f6578010557f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.826294 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v695\" (UniqueName: \"kubernetes.io/projected/5a10ac5e-a435-45ef-b143-f6578010557f-kube-api-access-4v695\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.826512 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a10ac5e-a435-45ef-b143-f6578010557f-config\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.826544 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5a10ac5e-a435-45ef-b143-f6578010557f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.826631 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpfk5\" (UniqueName: \"kubernetes.io/projected/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-kube-api-access-cpfk5\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.826641 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.831895 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6f57f2e3-1d7a-4e6c-a67d-b8c81823310b" (UID: "6f57f2e3-1d7a-4e6c-a67d-b8c81823310b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.869097 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-config" (OuterVolumeSpecName: "config") pod "6f57f2e3-1d7a-4e6c-a67d-b8c81823310b" (UID: "6f57f2e3-1d7a-4e6c-a67d-b8c81823310b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.912881 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.912929 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.937301 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a10ac5e-a435-45ef-b143-f6578010557f-config\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.937401 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5a10ac5e-a435-45ef-b143-f6578010557f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.937468 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a10ac5e-a435-45ef-b143-f6578010557f-scripts\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.937499 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a10ac5e-a435-45ef-b143-f6578010557f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.937572 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a10ac5e-a435-45ef-b143-f6578010557f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.937609 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a10ac5e-a435-45ef-b143-f6578010557f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.937633 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v695\" (UniqueName: \"kubernetes.io/projected/5a10ac5e-a435-45ef-b143-f6578010557f-kube-api-access-4v695\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.937693 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.937707 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.939038 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a10ac5e-a435-45ef-b143-f6578010557f-config\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.939313 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5a10ac5e-a435-45ef-b143-f6578010557f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.939823 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a10ac5e-a435-45ef-b143-f6578010557f-scripts\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.948641 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a10ac5e-a435-45ef-b143-f6578010557f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.962012 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a10ac5e-a435-45ef-b143-f6578010557f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.974171 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v695\" (UniqueName: \"kubernetes.io/projected/5a10ac5e-a435-45ef-b143-f6578010557f-kube-api-access-4v695\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:12 crc kubenswrapper[4806]: I1127 10:37:12.975038 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a10ac5e-a435-45ef-b143-f6578010557f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"5a10ac5e-a435-45ef-b143-f6578010557f\") " pod="openstack/ovn-northd-0" Nov 27 10:37:13 crc kubenswrapper[4806]: I1127 10:37:13.106291 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 27 10:37:13 crc kubenswrapper[4806]: I1127 10:37:13.154878 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 27 10:37:13 crc kubenswrapper[4806]: I1127 10:37:13.343745 4806 generic.go:334] "Generic (PLEG): container finished" podID="9aec6519-44e3-499c-bdce-151bf8a13a25" containerID="953b0dc7f1e2e636ab4ac6979a43ca6c14266fc05e6a46a0f7bfaaf994379a72" exitCode=0 Nov 27 10:37:13 crc kubenswrapper[4806]: I1127 10:37:13.343963 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-9hn6t" event={"ID":"9aec6519-44e3-499c-bdce-151bf8a13a25","Type":"ContainerDied","Data":"953b0dc7f1e2e636ab4ac6979a43ca6c14266fc05e6a46a0f7bfaaf994379a72"} Nov 27 10:37:13 crc kubenswrapper[4806]: I1127 10:37:13.358156 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-xg9f9" event={"ID":"a81c0343-88d6-42cf-bc8f-5a3b8722a137","Type":"ContainerStarted","Data":"73841630663bce7130b5bddb31dbba84bad2928b9704d8c66331b928e7314605"} Nov 27 10:37:13 crc kubenswrapper[4806]: I1127 10:37:13.398817 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-672fg" Nov 27 10:37:13 crc kubenswrapper[4806]: I1127 10:37:13.423269 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-672fg" event={"ID":"6f57f2e3-1d7a-4e6c-a67d-b8c81823310b","Type":"ContainerDied","Data":"fe3f0e2635539cb8cf82dc1a2a76b8aa4750f66265359e769b6fa0a81894b5dd"} Nov 27 10:37:13 crc kubenswrapper[4806]: I1127 10:37:13.423400 4806 scope.go:117] "RemoveContainer" containerID="ff59c59ea0bf11c03fd13fc074f34fca6311a75a31cb0c7735cb9bff2bb4f834" Nov 27 10:37:13 crc kubenswrapper[4806]: I1127 10:37:13.478101 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-xg9f9" podStartSLOduration=3.478077057 podStartE2EDuration="3.478077057s" podCreationTimestamp="2025-11-27 10:37:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:37:13.450865972 +0000 UTC m=+938.037456736" watchObservedRunningTime="2025-11-27 10:37:13.478077057 +0000 UTC m=+938.064667821" Nov 27 10:37:13 crc kubenswrapper[4806]: I1127 10:37:13.596669 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-672fg"] Nov 27 10:37:13 crc kubenswrapper[4806]: I1127 10:37:13.624650 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-672fg"] Nov 27 10:37:13 crc kubenswrapper[4806]: I1127 10:37:13.634321 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 27 10:37:13 crc kubenswrapper[4806]: I1127 10:37:13.671359 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.144684 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f57f2e3-1d7a-4e6c-a67d-b8c81823310b" path="/var/lib/kubelet/pods/6f57f2e3-1d7a-4e6c-a67d-b8c81823310b/volumes" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.264697 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.265624 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.402664 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-w7d78"] Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.405425 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-w7d78" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.411748 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-w7d78"] Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.417143 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-9hn6t" event={"ID":"9aec6519-44e3-499c-bdce-151bf8a13a25","Type":"ContainerStarted","Data":"897eb0bb82fed8ef6c3e6470f0373b2a88c89dc8027e91bf6b052f5bc063db6a"} Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.417756 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.423156 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5a10ac5e-a435-45ef-b143-f6578010557f","Type":"ContainerStarted","Data":"3d6219923e32cb3a7fc4d214816898565ce9eb93d580b19f8140c2cd88ff4777"} Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.447873 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.452320 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.452397 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.452469 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.453369 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc655ac750298c594d4faac6cef9782fc2d5a20661ea69277b9fafe78a2e81f0"} pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.453433 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" containerID="cri-o://dc655ac750298c594d4faac6cef9782fc2d5a20661ea69277b9fafe78a2e81f0" gracePeriod=600 Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.469009 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-4a42-account-create-update-rm55v"] Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.470779 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4a42-account-create-update-rm55v" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.478604 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.487508 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99b8v\" (UniqueName: \"kubernetes.io/projected/ff5722e8-f491-4a02-836a-f2953334ec7d-kube-api-access-99b8v\") pod \"keystone-db-create-w7d78\" (UID: \"ff5722e8-f491-4a02-836a-f2953334ec7d\") " pod="openstack/keystone-db-create-w7d78" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.487651 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff5722e8-f491-4a02-836a-f2953334ec7d-operator-scripts\") pod \"keystone-db-create-w7d78\" (UID: \"ff5722e8-f491-4a02-836a-f2953334ec7d\") " pod="openstack/keystone-db-create-w7d78" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.492550 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4a42-account-create-update-rm55v"] Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.499927 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-9hn6t" podStartSLOduration=3.499898764 podStartE2EDuration="3.499898764s" podCreationTimestamp="2025-11-27 10:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:37:14.476598847 +0000 UTC m=+939.063189611" watchObservedRunningTime="2025-11-27 10:37:14.499898764 +0000 UTC m=+939.086489528" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.603388 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff5722e8-f491-4a02-836a-f2953334ec7d-operator-scripts\") pod \"keystone-db-create-w7d78\" (UID: \"ff5722e8-f491-4a02-836a-f2953334ec7d\") " pod="openstack/keystone-db-create-w7d78" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.603523 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1020f922-5226-4961-aa56-04c97874b6a6-operator-scripts\") pod \"keystone-4a42-account-create-update-rm55v\" (UID: \"1020f922-5226-4961-aa56-04c97874b6a6\") " pod="openstack/keystone-4a42-account-create-update-rm55v" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.603644 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99b8v\" (UniqueName: \"kubernetes.io/projected/ff5722e8-f491-4a02-836a-f2953334ec7d-kube-api-access-99b8v\") pod \"keystone-db-create-w7d78\" (UID: \"ff5722e8-f491-4a02-836a-f2953334ec7d\") " pod="openstack/keystone-db-create-w7d78" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.603699 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwmqh\" (UniqueName: \"kubernetes.io/projected/1020f922-5226-4961-aa56-04c97874b6a6-kube-api-access-mwmqh\") pod \"keystone-4a42-account-create-update-rm55v\" (UID: \"1020f922-5226-4961-aa56-04c97874b6a6\") " pod="openstack/keystone-4a42-account-create-update-rm55v" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.605376 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff5722e8-f491-4a02-836a-f2953334ec7d-operator-scripts\") pod \"keystone-db-create-w7d78\" (UID: \"ff5722e8-f491-4a02-836a-f2953334ec7d\") " pod="openstack/keystone-db-create-w7d78" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.654940 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99b8v\" (UniqueName: \"kubernetes.io/projected/ff5722e8-f491-4a02-836a-f2953334ec7d-kube-api-access-99b8v\") pod \"keystone-db-create-w7d78\" (UID: \"ff5722e8-f491-4a02-836a-f2953334ec7d\") " pod="openstack/keystone-db-create-w7d78" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.680531 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-lkmxz"] Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.694329 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-lkmxz" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.707940 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwmqh\" (UniqueName: \"kubernetes.io/projected/1020f922-5226-4961-aa56-04c97874b6a6-kube-api-access-mwmqh\") pod \"keystone-4a42-account-create-update-rm55v\" (UID: \"1020f922-5226-4961-aa56-04c97874b6a6\") " pod="openstack/keystone-4a42-account-create-update-rm55v" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.713108 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1020f922-5226-4961-aa56-04c97874b6a6-operator-scripts\") pod \"keystone-4a42-account-create-update-rm55v\" (UID: \"1020f922-5226-4961-aa56-04c97874b6a6\") " pod="openstack/keystone-4a42-account-create-update-rm55v" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.713989 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1020f922-5226-4961-aa56-04c97874b6a6-operator-scripts\") pod \"keystone-4a42-account-create-update-rm55v\" (UID: \"1020f922-5226-4961-aa56-04c97874b6a6\") " pod="openstack/keystone-4a42-account-create-update-rm55v" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.727531 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-lkmxz"] Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.734658 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-w7d78" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.780466 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwmqh\" (UniqueName: \"kubernetes.io/projected/1020f922-5226-4961-aa56-04c97874b6a6-kube-api-access-mwmqh\") pod \"keystone-4a42-account-create-update-rm55v\" (UID: \"1020f922-5226-4961-aa56-04c97874b6a6\") " pod="openstack/keystone-4a42-account-create-update-rm55v" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.814110 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4a42-account-create-update-rm55v" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.814862 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b92e4ceb-87fe-4521-b931-9e792f9e4590-operator-scripts\") pod \"placement-db-create-lkmxz\" (UID: \"b92e4ceb-87fe-4521-b931-9e792f9e4590\") " pod="openstack/placement-db-create-lkmxz" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.814955 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfkvt\" (UniqueName: \"kubernetes.io/projected/b92e4ceb-87fe-4521-b931-9e792f9e4590-kube-api-access-gfkvt\") pod \"placement-db-create-lkmxz\" (UID: \"b92e4ceb-87fe-4521-b931-9e792f9e4590\") " pod="openstack/placement-db-create-lkmxz" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.826912 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-3292-account-create-update-p4vfd"] Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.828224 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3292-account-create-update-p4vfd" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.832258 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.839977 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3292-account-create-update-p4vfd"] Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.917931 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/685cee37-d40c-4ee8-ad87-c67acb12fd60-operator-scripts\") pod \"placement-3292-account-create-update-p4vfd\" (UID: \"685cee37-d40c-4ee8-ad87-c67acb12fd60\") " pod="openstack/placement-3292-account-create-update-p4vfd" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.918059 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b92e4ceb-87fe-4521-b931-9e792f9e4590-operator-scripts\") pod \"placement-db-create-lkmxz\" (UID: \"b92e4ceb-87fe-4521-b931-9e792f9e4590\") " pod="openstack/placement-db-create-lkmxz" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.918120 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfkvt\" (UniqueName: \"kubernetes.io/projected/b92e4ceb-87fe-4521-b931-9e792f9e4590-kube-api-access-gfkvt\") pod \"placement-db-create-lkmxz\" (UID: \"b92e4ceb-87fe-4521-b931-9e792f9e4590\") " pod="openstack/placement-db-create-lkmxz" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.918147 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgr2v\" (UniqueName: \"kubernetes.io/projected/685cee37-d40c-4ee8-ad87-c67acb12fd60-kube-api-access-hgr2v\") pod \"placement-3292-account-create-update-p4vfd\" (UID: \"685cee37-d40c-4ee8-ad87-c67acb12fd60\") " pod="openstack/placement-3292-account-create-update-p4vfd" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.919144 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b92e4ceb-87fe-4521-b931-9e792f9e4590-operator-scripts\") pod \"placement-db-create-lkmxz\" (UID: \"b92e4ceb-87fe-4521-b931-9e792f9e4590\") " pod="openstack/placement-db-create-lkmxz" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.923417 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-qfb6s"] Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.924755 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-qfb6s" Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.938481 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-qfb6s"] Nov 27 10:37:14 crc kubenswrapper[4806]: I1127 10:37:14.951480 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfkvt\" (UniqueName: \"kubernetes.io/projected/b92e4ceb-87fe-4521-b931-9e792f9e4590-kube-api-access-gfkvt\") pod \"placement-db-create-lkmxz\" (UID: \"b92e4ceb-87fe-4521-b931-9e792f9e4590\") " pod="openstack/placement-db-create-lkmxz" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.018424 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-c11b-account-create-update-dkh64"] Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.020407 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc617567-5421-4521-8957-4b11667db29e-operator-scripts\") pod \"glance-db-create-qfb6s\" (UID: \"fc617567-5421-4521-8957-4b11667db29e\") " pod="openstack/glance-db-create-qfb6s" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.020460 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/685cee37-d40c-4ee8-ad87-c67acb12fd60-operator-scripts\") pod \"placement-3292-account-create-update-p4vfd\" (UID: \"685cee37-d40c-4ee8-ad87-c67acb12fd60\") " pod="openstack/placement-3292-account-create-update-p4vfd" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.020507 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rptmt\" (UniqueName: \"kubernetes.io/projected/fc617567-5421-4521-8957-4b11667db29e-kube-api-access-rptmt\") pod \"glance-db-create-qfb6s\" (UID: \"fc617567-5421-4521-8957-4b11667db29e\") " pod="openstack/glance-db-create-qfb6s" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.020572 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgr2v\" (UniqueName: \"kubernetes.io/projected/685cee37-d40c-4ee8-ad87-c67acb12fd60-kube-api-access-hgr2v\") pod \"placement-3292-account-create-update-p4vfd\" (UID: \"685cee37-d40c-4ee8-ad87-c67acb12fd60\") " pod="openstack/placement-3292-account-create-update-p4vfd" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.021810 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/685cee37-d40c-4ee8-ad87-c67acb12fd60-operator-scripts\") pod \"placement-3292-account-create-update-p4vfd\" (UID: \"685cee37-d40c-4ee8-ad87-c67acb12fd60\") " pod="openstack/placement-3292-account-create-update-p4vfd" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.029704 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c11b-account-create-update-dkh64"] Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.029828 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c11b-account-create-update-dkh64" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.033795 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.038767 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgr2v\" (UniqueName: \"kubernetes.io/projected/685cee37-d40c-4ee8-ad87-c67acb12fd60-kube-api-access-hgr2v\") pod \"placement-3292-account-create-update-p4vfd\" (UID: \"685cee37-d40c-4ee8-ad87-c67acb12fd60\") " pod="openstack/placement-3292-account-create-update-p4vfd" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.075086 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-lkmxz" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.124963 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3572cba-65ba-47eb-a98a-ff0c4b3feec8-operator-scripts\") pod \"glance-c11b-account-create-update-dkh64\" (UID: \"a3572cba-65ba-47eb-a98a-ff0c4b3feec8\") " pod="openstack/glance-c11b-account-create-update-dkh64" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.125061 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc617567-5421-4521-8957-4b11667db29e-operator-scripts\") pod \"glance-db-create-qfb6s\" (UID: \"fc617567-5421-4521-8957-4b11667db29e\") " pod="openstack/glance-db-create-qfb6s" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.125105 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rptmt\" (UniqueName: \"kubernetes.io/projected/fc617567-5421-4521-8957-4b11667db29e-kube-api-access-rptmt\") pod \"glance-db-create-qfb6s\" (UID: \"fc617567-5421-4521-8957-4b11667db29e\") " pod="openstack/glance-db-create-qfb6s" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.125182 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8rpq\" (UniqueName: \"kubernetes.io/projected/a3572cba-65ba-47eb-a98a-ff0c4b3feec8-kube-api-access-d8rpq\") pod \"glance-c11b-account-create-update-dkh64\" (UID: \"a3572cba-65ba-47eb-a98a-ff0c4b3feec8\") " pod="openstack/glance-c11b-account-create-update-dkh64" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.126010 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc617567-5421-4521-8957-4b11667db29e-operator-scripts\") pod \"glance-db-create-qfb6s\" (UID: \"fc617567-5421-4521-8957-4b11667db29e\") " pod="openstack/glance-db-create-qfb6s" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.148605 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rptmt\" (UniqueName: \"kubernetes.io/projected/fc617567-5421-4521-8957-4b11667db29e-kube-api-access-rptmt\") pod \"glance-db-create-qfb6s\" (UID: \"fc617567-5421-4521-8957-4b11667db29e\") " pod="openstack/glance-db-create-qfb6s" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.168754 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3292-account-create-update-p4vfd" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.227150 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8rpq\" (UniqueName: \"kubernetes.io/projected/a3572cba-65ba-47eb-a98a-ff0c4b3feec8-kube-api-access-d8rpq\") pod \"glance-c11b-account-create-update-dkh64\" (UID: \"a3572cba-65ba-47eb-a98a-ff0c4b3feec8\") " pod="openstack/glance-c11b-account-create-update-dkh64" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.227399 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3572cba-65ba-47eb-a98a-ff0c4b3feec8-operator-scripts\") pod \"glance-c11b-account-create-update-dkh64\" (UID: \"a3572cba-65ba-47eb-a98a-ff0c4b3feec8\") " pod="openstack/glance-c11b-account-create-update-dkh64" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.233445 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3572cba-65ba-47eb-a98a-ff0c4b3feec8-operator-scripts\") pod \"glance-c11b-account-create-update-dkh64\" (UID: \"a3572cba-65ba-47eb-a98a-ff0c4b3feec8\") " pod="openstack/glance-c11b-account-create-update-dkh64" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.248151 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8rpq\" (UniqueName: \"kubernetes.io/projected/a3572cba-65ba-47eb-a98a-ff0c4b3feec8-kube-api-access-d8rpq\") pod \"glance-c11b-account-create-update-dkh64\" (UID: \"a3572cba-65ba-47eb-a98a-ff0c4b3feec8\") " pod="openstack/glance-c11b-account-create-update-dkh64" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.249346 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-qfb6s" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.345825 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c11b-account-create-update-dkh64" Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.493883 4806 generic.go:334] "Generic (PLEG): container finished" podID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerID="dc655ac750298c594d4faac6cef9782fc2d5a20661ea69277b9fafe78a2e81f0" exitCode=0 Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.494265 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerDied","Data":"dc655ac750298c594d4faac6cef9782fc2d5a20661ea69277b9fafe78a2e81f0"} Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.494323 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"5908b1682df3b6a3dc72ea1b0d3756efb806c890ef007bbc5b48b3a697274cc4"} Nov 27 10:37:15 crc kubenswrapper[4806]: I1127 10:37:15.494344 4806 scope.go:117] "RemoveContainer" containerID="5e15c67a8ce5b7fd1f5bca346e1724cf329b9d00aa70feed8824c46529a879e6" Nov 27 10:37:16 crc kubenswrapper[4806]: I1127 10:37:16.162068 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 27 10:37:16 crc kubenswrapper[4806]: I1127 10:37:16.359754 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-qfb6s"] Nov 27 10:37:16 crc kubenswrapper[4806]: I1127 10:37:16.418472 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-w7d78"] Nov 27 10:37:16 crc kubenswrapper[4806]: I1127 10:37:16.514618 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-qfb6s" event={"ID":"fc617567-5421-4521-8957-4b11667db29e","Type":"ContainerStarted","Data":"c3505246598bc3722b0098b5562b1a29f61474c89f602660363022cba64b73ce"} Nov 27 10:37:16 crc kubenswrapper[4806]: I1127 10:37:16.519803 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5a10ac5e-a435-45ef-b143-f6578010557f","Type":"ContainerStarted","Data":"671b544021aaf48179380080e3e0cf34caaab45aad0d120b15a7929ac66cc958"} Nov 27 10:37:16 crc kubenswrapper[4806]: I1127 10:37:16.522978 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-w7d78" event={"ID":"ff5722e8-f491-4a02-836a-f2953334ec7d","Type":"ContainerStarted","Data":"16b73dfdde2d03a79c07adb11fe17d1751022f3cbb67f080233671c391df636f"} Nov 27 10:37:16 crc kubenswrapper[4806]: I1127 10:37:16.722266 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3292-account-create-update-p4vfd"] Nov 27 10:37:16 crc kubenswrapper[4806]: I1127 10:37:16.733640 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-lkmxz"] Nov 27 10:37:16 crc kubenswrapper[4806]: W1127 10:37:16.756905 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb92e4ceb_87fe_4521_b931_9e792f9e4590.slice/crio-d6601d959eeb37a1edd50c67cd4ffc6cb9fff8ce4acd6e6d87a52adc73077238 WatchSource:0}: Error finding container d6601d959eeb37a1edd50c67cd4ffc6cb9fff8ce4acd6e6d87a52adc73077238: Status 404 returned error can't find the container with id d6601d959eeb37a1edd50c67cd4ffc6cb9fff8ce4acd6e6d87a52adc73077238 Nov 27 10:37:16 crc kubenswrapper[4806]: I1127 10:37:16.829569 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4a42-account-create-update-rm55v"] Nov 27 10:37:16 crc kubenswrapper[4806]: I1127 10:37:16.982217 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-c11b-account-create-update-dkh64"] Nov 27 10:37:16 crc kubenswrapper[4806]: W1127 10:37:16.989894 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3572cba_65ba_47eb_a98a_ff0c4b3feec8.slice/crio-52bc445a4a17c793ca1ad07bd10f7e917da055035f3997bcd0e056a87acc42f3 WatchSource:0}: Error finding container 52bc445a4a17c793ca1ad07bd10f7e917da055035f3997bcd0e056a87acc42f3: Status 404 returned error can't find the container with id 52bc445a4a17c793ca1ad07bd10f7e917da055035f3997bcd0e056a87acc42f3 Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.536085 4806 generic.go:334] "Generic (PLEG): container finished" podID="1020f922-5226-4961-aa56-04c97874b6a6" containerID="d6b07eb372a44b87726e1e2e0264485b33b0c4773996c678d1f60e99f2f60d6a" exitCode=0 Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.536148 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4a42-account-create-update-rm55v" event={"ID":"1020f922-5226-4961-aa56-04c97874b6a6","Type":"ContainerDied","Data":"d6b07eb372a44b87726e1e2e0264485b33b0c4773996c678d1f60e99f2f60d6a"} Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.536512 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4a42-account-create-update-rm55v" event={"ID":"1020f922-5226-4961-aa56-04c97874b6a6","Type":"ContainerStarted","Data":"31b65e46db6105d1e34da39bfbe83cbaec4a223dafaf3b11fab977ae716a71b5"} Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.540006 4806 generic.go:334] "Generic (PLEG): container finished" podID="685cee37-d40c-4ee8-ad87-c67acb12fd60" containerID="cf7772a86778265847fd4a67a69bd0f1b2072cd2f2fee80e3b9d73606dd7a0d6" exitCode=0 Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.540071 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3292-account-create-update-p4vfd" event={"ID":"685cee37-d40c-4ee8-ad87-c67acb12fd60","Type":"ContainerDied","Data":"cf7772a86778265847fd4a67a69bd0f1b2072cd2f2fee80e3b9d73606dd7a0d6"} Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.540097 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3292-account-create-update-p4vfd" event={"ID":"685cee37-d40c-4ee8-ad87-c67acb12fd60","Type":"ContainerStarted","Data":"2415d1f8bc93c5e9cbd5c33c71c8853c6bbf21f3513878bb59d2c497c529e25e"} Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.542319 4806 generic.go:334] "Generic (PLEG): container finished" podID="b92e4ceb-87fe-4521-b931-9e792f9e4590" containerID="2abd910320dc8930648b10301139f8fa31ceee0dca10e90166d2952768870271" exitCode=0 Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.542395 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-lkmxz" event={"ID":"b92e4ceb-87fe-4521-b931-9e792f9e4590","Type":"ContainerDied","Data":"2abd910320dc8930648b10301139f8fa31ceee0dca10e90166d2952768870271"} Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.542426 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-lkmxz" event={"ID":"b92e4ceb-87fe-4521-b931-9e792f9e4590","Type":"ContainerStarted","Data":"d6601d959eeb37a1edd50c67cd4ffc6cb9fff8ce4acd6e6d87a52adc73077238"} Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.544372 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5a10ac5e-a435-45ef-b143-f6578010557f","Type":"ContainerStarted","Data":"4e9382a457e24497f87b346d9f37d8c838329f64d6fe28823c7b6f442f40b316"} Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.545316 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.549150 4806 generic.go:334] "Generic (PLEG): container finished" podID="ff5722e8-f491-4a02-836a-f2953334ec7d" containerID="83a572048feb1595316cc2d16cd603da84a3d7b676e9d80a2f4f7c58922bbcf1" exitCode=0 Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.549208 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-w7d78" event={"ID":"ff5722e8-f491-4a02-836a-f2953334ec7d","Type":"ContainerDied","Data":"83a572048feb1595316cc2d16cd603da84a3d7b676e9d80a2f4f7c58922bbcf1"} Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.557684 4806 generic.go:334] "Generic (PLEG): container finished" podID="fc617567-5421-4521-8957-4b11667db29e" containerID="f3ab201c5083e96c309a8526d027d7add5ab4d4266af330b19b20b947a743221" exitCode=0 Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.557927 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-qfb6s" event={"ID":"fc617567-5421-4521-8957-4b11667db29e","Type":"ContainerDied","Data":"f3ab201c5083e96c309a8526d027d7add5ab4d4266af330b19b20b947a743221"} Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.565366 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c11b-account-create-update-dkh64" event={"ID":"a3572cba-65ba-47eb-a98a-ff0c4b3feec8","Type":"ContainerStarted","Data":"b3bb5f91d43d19f6b89a5a7e99071c2ba7ecca58bdf586c7c400bb174c06088e"} Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.565437 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c11b-account-create-update-dkh64" event={"ID":"a3572cba-65ba-47eb-a98a-ff0c4b3feec8","Type":"ContainerStarted","Data":"52bc445a4a17c793ca1ad07bd10f7e917da055035f3997bcd0e056a87acc42f3"} Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.632039 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.531521992 podStartE2EDuration="5.632006001s" podCreationTimestamp="2025-11-27 10:37:12 +0000 UTC" firstStartedPulling="2025-11-27 10:37:13.64945829 +0000 UTC m=+938.236049054" lastFinishedPulling="2025-11-27 10:37:15.749942299 +0000 UTC m=+940.336533063" observedRunningTime="2025-11-27 10:37:17.630865459 +0000 UTC m=+942.217456223" watchObservedRunningTime="2025-11-27 10:37:17.632006001 +0000 UTC m=+942.218596765" Nov 27 10:37:17 crc kubenswrapper[4806]: I1127 10:37:17.654654 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-c11b-account-create-update-dkh64" podStartSLOduration=3.654614338 podStartE2EDuration="3.654614338s" podCreationTimestamp="2025-11-27 10:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:37:17.651966264 +0000 UTC m=+942.238557028" watchObservedRunningTime="2025-11-27 10:37:17.654614338 +0000 UTC m=+942.241205112" Nov 27 10:37:18 crc kubenswrapper[4806]: I1127 10:37:18.577836 4806 generic.go:334] "Generic (PLEG): container finished" podID="a3572cba-65ba-47eb-a98a-ff0c4b3feec8" containerID="b3bb5f91d43d19f6b89a5a7e99071c2ba7ecca58bdf586c7c400bb174c06088e" exitCode=0 Nov 27 10:37:18 crc kubenswrapper[4806]: I1127 10:37:18.578021 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c11b-account-create-update-dkh64" event={"ID":"a3572cba-65ba-47eb-a98a-ff0c4b3feec8","Type":"ContainerDied","Data":"b3bb5f91d43d19f6b89a5a7e99071c2ba7ecca58bdf586c7c400bb174c06088e"} Nov 27 10:37:20 crc kubenswrapper[4806]: I1127 10:37:20.598142 4806 generic.go:334] "Generic (PLEG): container finished" podID="954f26d7-2bdc-45e3-a07b-d9924469f0a1" containerID="96a9494ee6564f52a58e01b832a4f8f8d3d7c39e07840c0adfcef6d9414c5d1b" exitCode=0 Nov 27 10:37:20 crc kubenswrapper[4806]: I1127 10:37:20.599105 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"954f26d7-2bdc-45e3-a07b-d9924469f0a1","Type":"ContainerDied","Data":"96a9494ee6564f52a58e01b832a4f8f8d3d7c39e07840c0adfcef6d9414c5d1b"} Nov 27 10:37:20 crc kubenswrapper[4806]: I1127 10:37:20.606087 4806 generic.go:334] "Generic (PLEG): container finished" podID="08137128-e0d1-49d2-abf1-1242c1cb58ce" containerID="e7c3736208836a8b527976f788183343ce7a097c822f342b20cd72859c3b83dd" exitCode=0 Nov 27 10:37:20 crc kubenswrapper[4806]: I1127 10:37:20.606129 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"08137128-e0d1-49d2-abf1-1242c1cb58ce","Type":"ContainerDied","Data":"e7c3736208836a8b527976f788183343ce7a097c822f342b20cd72859c3b83dd"} Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.027664 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ckhq4"] Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.034266 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.055403 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ckhq4"] Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.088731 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4b050be-219f-4adc-83b5-52da68b7bade-utilities\") pod \"certified-operators-ckhq4\" (UID: \"f4b050be-219f-4adc-83b5-52da68b7bade\") " pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.088796 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hflm\" (UniqueName: \"kubernetes.io/projected/f4b050be-219f-4adc-83b5-52da68b7bade-kube-api-access-7hflm\") pod \"certified-operators-ckhq4\" (UID: \"f4b050be-219f-4adc-83b5-52da68b7bade\") " pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.088875 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4b050be-219f-4adc-83b5-52da68b7bade-catalog-content\") pod \"certified-operators-ckhq4\" (UID: \"f4b050be-219f-4adc-83b5-52da68b7bade\") " pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.190754 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hflm\" (UniqueName: \"kubernetes.io/projected/f4b050be-219f-4adc-83b5-52da68b7bade-kube-api-access-7hflm\") pod \"certified-operators-ckhq4\" (UID: \"f4b050be-219f-4adc-83b5-52da68b7bade\") " pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.190850 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4b050be-219f-4adc-83b5-52da68b7bade-catalog-content\") pod \"certified-operators-ckhq4\" (UID: \"f4b050be-219f-4adc-83b5-52da68b7bade\") " pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.190948 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4b050be-219f-4adc-83b5-52da68b7bade-utilities\") pod \"certified-operators-ckhq4\" (UID: \"f4b050be-219f-4adc-83b5-52da68b7bade\") " pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.191380 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4b050be-219f-4adc-83b5-52da68b7bade-utilities\") pod \"certified-operators-ckhq4\" (UID: \"f4b050be-219f-4adc-83b5-52da68b7bade\") " pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.192427 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4b050be-219f-4adc-83b5-52da68b7bade-catalog-content\") pod \"certified-operators-ckhq4\" (UID: \"f4b050be-219f-4adc-83b5-52da68b7bade\") " pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.212549 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hflm\" (UniqueName: \"kubernetes.io/projected/f4b050be-219f-4adc-83b5-52da68b7bade-kube-api-access-7hflm\") pod \"certified-operators-ckhq4\" (UID: \"f4b050be-219f-4adc-83b5-52da68b7bade\") " pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.356190 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.584905 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-w7d78" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.617843 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-qfb6s" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.639199 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3292-account-create-update-p4vfd" event={"ID":"685cee37-d40c-4ee8-ad87-c67acb12fd60","Type":"ContainerDied","Data":"2415d1f8bc93c5e9cbd5c33c71c8853c6bbf21f3513878bb59d2c497c529e25e"} Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.639271 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2415d1f8bc93c5e9cbd5c33c71c8853c6bbf21f3513878bb59d2c497c529e25e" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.639390 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3292-account-create-update-p4vfd" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.667698 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-lkmxz" event={"ID":"b92e4ceb-87fe-4521-b931-9e792f9e4590","Type":"ContainerDied","Data":"d6601d959eeb37a1edd50c67cd4ffc6cb9fff8ce4acd6e6d87a52adc73077238"} Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.667753 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6601d959eeb37a1edd50c67cd4ffc6cb9fff8ce4acd6e6d87a52adc73077238" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.686450 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.706189 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c11b-account-create-update-dkh64" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.707916 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/685cee37-d40c-4ee8-ad87-c67acb12fd60-operator-scripts\") pod \"685cee37-d40c-4ee8-ad87-c67acb12fd60\" (UID: \"685cee37-d40c-4ee8-ad87-c67acb12fd60\") " Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.707945 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99b8v\" (UniqueName: \"kubernetes.io/projected/ff5722e8-f491-4a02-836a-f2953334ec7d-kube-api-access-99b8v\") pod \"ff5722e8-f491-4a02-836a-f2953334ec7d\" (UID: \"ff5722e8-f491-4a02-836a-f2953334ec7d\") " Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.708046 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgr2v\" (UniqueName: \"kubernetes.io/projected/685cee37-d40c-4ee8-ad87-c67acb12fd60-kube-api-access-hgr2v\") pod \"685cee37-d40c-4ee8-ad87-c67acb12fd60\" (UID: \"685cee37-d40c-4ee8-ad87-c67acb12fd60\") " Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.708104 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff5722e8-f491-4a02-836a-f2953334ec7d-operator-scripts\") pod \"ff5722e8-f491-4a02-836a-f2953334ec7d\" (UID: \"ff5722e8-f491-4a02-836a-f2953334ec7d\") " Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.708141 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rptmt\" (UniqueName: \"kubernetes.io/projected/fc617567-5421-4521-8957-4b11667db29e-kube-api-access-rptmt\") pod \"fc617567-5421-4521-8957-4b11667db29e\" (UID: \"fc617567-5421-4521-8957-4b11667db29e\") " Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.708223 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc617567-5421-4521-8957-4b11667db29e-operator-scripts\") pod \"fc617567-5421-4521-8957-4b11667db29e\" (UID: \"fc617567-5421-4521-8957-4b11667db29e\") " Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.709886 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc617567-5421-4521-8957-4b11667db29e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fc617567-5421-4521-8957-4b11667db29e" (UID: "fc617567-5421-4521-8957-4b11667db29e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.710645 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5722e8-f491-4a02-836a-f2953334ec7d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ff5722e8-f491-4a02-836a-f2953334ec7d" (UID: "ff5722e8-f491-4a02-836a-f2953334ec7d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.712328 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-w7d78" event={"ID":"ff5722e8-f491-4a02-836a-f2953334ec7d","Type":"ContainerDied","Data":"16b73dfdde2d03a79c07adb11fe17d1751022f3cbb67f080233671c391df636f"} Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.712383 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16b73dfdde2d03a79c07adb11fe17d1751022f3cbb67f080233671c391df636f" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.712481 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-w7d78" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.717123 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/685cee37-d40c-4ee8-ad87-c67acb12fd60-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "685cee37-d40c-4ee8-ad87-c67acb12fd60" (UID: "685cee37-d40c-4ee8-ad87-c67acb12fd60"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.718275 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4a42-account-create-update-rm55v" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.721589 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc617567-5421-4521-8957-4b11667db29e-kube-api-access-rptmt" (OuterVolumeSpecName: "kube-api-access-rptmt") pod "fc617567-5421-4521-8957-4b11667db29e" (UID: "fc617567-5421-4521-8957-4b11667db29e"). InnerVolumeSpecName "kube-api-access-rptmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.730561 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/685cee37-d40c-4ee8-ad87-c67acb12fd60-kube-api-access-hgr2v" (OuterVolumeSpecName: "kube-api-access-hgr2v") pod "685cee37-d40c-4ee8-ad87-c67acb12fd60" (UID: "685cee37-d40c-4ee8-ad87-c67acb12fd60"). InnerVolumeSpecName "kube-api-access-hgr2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.735571 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff5722e8-f491-4a02-836a-f2953334ec7d-kube-api-access-99b8v" (OuterVolumeSpecName: "kube-api-access-99b8v") pod "ff5722e8-f491-4a02-836a-f2953334ec7d" (UID: "ff5722e8-f491-4a02-836a-f2953334ec7d"). InnerVolumeSpecName "kube-api-access-99b8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.741649 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-qfb6s" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.745266 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-qfb6s" event={"ID":"fc617567-5421-4521-8957-4b11667db29e","Type":"ContainerDied","Data":"c3505246598bc3722b0098b5562b1a29f61474c89f602660363022cba64b73ce"} Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.745378 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3505246598bc3722b0098b5562b1a29f61474c89f602660363022cba64b73ce" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.766984 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-lkmxz" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.768228 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-c11b-account-create-update-dkh64" event={"ID":"a3572cba-65ba-47eb-a98a-ff0c4b3feec8","Type":"ContainerDied","Data":"52bc445a4a17c793ca1ad07bd10f7e917da055035f3997bcd0e056a87acc42f3"} Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.768292 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52bc445a4a17c793ca1ad07bd10f7e917da055035f3997bcd0e056a87acc42f3" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.773065 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-c11b-account-create-update-dkh64" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.775107 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4a42-account-create-update-rm55v" event={"ID":"1020f922-5226-4961-aa56-04c97874b6a6","Type":"ContainerDied","Data":"31b65e46db6105d1e34da39bfbe83cbaec4a223dafaf3b11fab977ae716a71b5"} Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.775162 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31b65e46db6105d1e34da39bfbe83cbaec4a223dafaf3b11fab977ae716a71b5" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.775253 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4a42-account-create-update-rm55v" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.810151 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3572cba-65ba-47eb-a98a-ff0c4b3feec8-operator-scripts\") pod \"a3572cba-65ba-47eb-a98a-ff0c4b3feec8\" (UID: \"a3572cba-65ba-47eb-a98a-ff0c4b3feec8\") " Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.810240 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b92e4ceb-87fe-4521-b931-9e792f9e4590-operator-scripts\") pod \"b92e4ceb-87fe-4521-b931-9e792f9e4590\" (UID: \"b92e4ceb-87fe-4521-b931-9e792f9e4590\") " Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.810286 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8rpq\" (UniqueName: \"kubernetes.io/projected/a3572cba-65ba-47eb-a98a-ff0c4b3feec8-kube-api-access-d8rpq\") pod \"a3572cba-65ba-47eb-a98a-ff0c4b3feec8\" (UID: \"a3572cba-65ba-47eb-a98a-ff0c4b3feec8\") " Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.810391 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfkvt\" (UniqueName: \"kubernetes.io/projected/b92e4ceb-87fe-4521-b931-9e792f9e4590-kube-api-access-gfkvt\") pod \"b92e4ceb-87fe-4521-b931-9e792f9e4590\" (UID: \"b92e4ceb-87fe-4521-b931-9e792f9e4590\") " Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.810442 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwmqh\" (UniqueName: \"kubernetes.io/projected/1020f922-5226-4961-aa56-04c97874b6a6-kube-api-access-mwmqh\") pod \"1020f922-5226-4961-aa56-04c97874b6a6\" (UID: \"1020f922-5226-4961-aa56-04c97874b6a6\") " Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.810486 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1020f922-5226-4961-aa56-04c97874b6a6-operator-scripts\") pod \"1020f922-5226-4961-aa56-04c97874b6a6\" (UID: \"1020f922-5226-4961-aa56-04c97874b6a6\") " Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.811270 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgr2v\" (UniqueName: \"kubernetes.io/projected/685cee37-d40c-4ee8-ad87-c67acb12fd60-kube-api-access-hgr2v\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.811296 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff5722e8-f491-4a02-836a-f2953334ec7d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.811308 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rptmt\" (UniqueName: \"kubernetes.io/projected/fc617567-5421-4521-8957-4b11667db29e-kube-api-access-rptmt\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.811320 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc617567-5421-4521-8957-4b11667db29e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.811330 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/685cee37-d40c-4ee8-ad87-c67acb12fd60-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.811342 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99b8v\" (UniqueName: \"kubernetes.io/projected/ff5722e8-f491-4a02-836a-f2953334ec7d-kube-api-access-99b8v\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.812372 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3572cba-65ba-47eb-a98a-ff0c4b3feec8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a3572cba-65ba-47eb-a98a-ff0c4b3feec8" (UID: "a3572cba-65ba-47eb-a98a-ff0c4b3feec8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.812932 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b92e4ceb-87fe-4521-b931-9e792f9e4590-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b92e4ceb-87fe-4521-b931-9e792f9e4590" (UID: "b92e4ceb-87fe-4521-b931-9e792f9e4590"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.816847 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1020f922-5226-4961-aa56-04c97874b6a6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1020f922-5226-4961-aa56-04c97874b6a6" (UID: "1020f922-5226-4961-aa56-04c97874b6a6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.822210 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b92e4ceb-87fe-4521-b931-9e792f9e4590-kube-api-access-gfkvt" (OuterVolumeSpecName: "kube-api-access-gfkvt") pod "b92e4ceb-87fe-4521-b931-9e792f9e4590" (UID: "b92e4ceb-87fe-4521-b931-9e792f9e4590"). InnerVolumeSpecName "kube-api-access-gfkvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.822597 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1020f922-5226-4961-aa56-04c97874b6a6-kube-api-access-mwmqh" (OuterVolumeSpecName: "kube-api-access-mwmqh") pod "1020f922-5226-4961-aa56-04c97874b6a6" (UID: "1020f922-5226-4961-aa56-04c97874b6a6"). InnerVolumeSpecName "kube-api-access-mwmqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.829883 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3572cba-65ba-47eb-a98a-ff0c4b3feec8-kube-api-access-d8rpq" (OuterVolumeSpecName: "kube-api-access-d8rpq") pod "a3572cba-65ba-47eb-a98a-ff0c4b3feec8" (UID: "a3572cba-65ba-47eb-a98a-ff0c4b3feec8"). InnerVolumeSpecName "kube-api-access-d8rpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.833433 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p2w2s"] Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.833863 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" podUID="87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7" containerName="dnsmasq-dns" containerID="cri-o://6d6f1bb081e4d9609ef5fc110a6d05da448158bf2d1c12676d23807fc8dd6094" gracePeriod=10 Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.913616 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8rpq\" (UniqueName: \"kubernetes.io/projected/a3572cba-65ba-47eb-a98a-ff0c4b3feec8-kube-api-access-d8rpq\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.913662 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfkvt\" (UniqueName: \"kubernetes.io/projected/b92e4ceb-87fe-4521-b931-9e792f9e4590-kube-api-access-gfkvt\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.913673 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwmqh\" (UniqueName: \"kubernetes.io/projected/1020f922-5226-4961-aa56-04c97874b6a6-kube-api-access-mwmqh\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.913682 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1020f922-5226-4961-aa56-04c97874b6a6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.913693 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3572cba-65ba-47eb-a98a-ff0c4b3feec8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:21 crc kubenswrapper[4806]: I1127 10:37:21.913703 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b92e4ceb-87fe-4521-b931-9e792f9e4590-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.410881 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ckhq4"] Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.790430 4806 generic.go:334] "Generic (PLEG): container finished" podID="87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7" containerID="6d6f1bb081e4d9609ef5fc110a6d05da448158bf2d1c12676d23807fc8dd6094" exitCode=0 Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.790514 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" event={"ID":"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7","Type":"ContainerDied","Data":"6d6f1bb081e4d9609ef5fc110a6d05da448158bf2d1c12676d23807fc8dd6094"} Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.800534 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"954f26d7-2bdc-45e3-a07b-d9924469f0a1","Type":"ContainerStarted","Data":"20acdf98b7f0a69423bc1346d4f34b766d96fc603ba847c5f00d8813173ee867"} Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.800872 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.815198 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"08137128-e0d1-49d2-abf1-1242c1cb58ce","Type":"ContainerStarted","Data":"fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea"} Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.815734 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.818894 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pc2bp" event={"ID":"a19e7111-3ef1-45a9-ae31-0d53a2a6615f","Type":"ContainerStarted","Data":"85bec44b35fe89b343437b8e2b5eacf64f5320da66e448ec00dd6effb530b151"} Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.824648 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3292-account-create-update-p4vfd" Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.825389 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ckhq4" event={"ID":"f4b050be-219f-4adc-83b5-52da68b7bade","Type":"ContainerStarted","Data":"2d63dbbfd7bce0dd94b46761eed7fd02d4e8d2cb0c8005147dd820a388ceb9a5"} Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.825421 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ckhq4" event={"ID":"f4b050be-219f-4adc-83b5-52da68b7bade","Type":"ContainerStarted","Data":"43950de7e7bc7fbaf8555ce23c6fe60f2fd40facf06f39e1176902affe651726"} Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.825462 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-lkmxz" Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.866100 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=51.908155283 podStartE2EDuration="1m12.866077447s" podCreationTimestamp="2025-11-27 10:36:10 +0000 UTC" firstStartedPulling="2025-11-27 10:36:26.502961417 +0000 UTC m=+891.089552181" lastFinishedPulling="2025-11-27 10:36:47.460883581 +0000 UTC m=+912.047474345" observedRunningTime="2025-11-27 10:37:22.861932461 +0000 UTC m=+947.448523235" watchObservedRunningTime="2025-11-27 10:37:22.866077447 +0000 UTC m=+947.452668211" Nov 27 10:37:22 crc kubenswrapper[4806]: I1127 10:37:22.959996 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=52.716723014 podStartE2EDuration="1m12.95997583s" podCreationTimestamp="2025-11-27 10:36:10 +0000 UTC" firstStartedPulling="2025-11-27 10:36:26.503307446 +0000 UTC m=+891.089898210" lastFinishedPulling="2025-11-27 10:36:46.746560262 +0000 UTC m=+911.333151026" observedRunningTime="2025-11-27 10:37:22.952967936 +0000 UTC m=+947.539558700" watchObservedRunningTime="2025-11-27 10:37:22.95997583 +0000 UTC m=+947.546566594" Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.027791 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.099600 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-config\") pod \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\" (UID: \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\") " Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.101592 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdr6s\" (UniqueName: \"kubernetes.io/projected/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-kube-api-access-hdr6s\") pod \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\" (UID: \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\") " Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.101999 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-dns-svc\") pod \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\" (UID: \"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7\") " Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.107081 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-kube-api-access-hdr6s" (OuterVolumeSpecName: "kube-api-access-hdr6s") pod "87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7" (UID: "87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7"). InnerVolumeSpecName "kube-api-access-hdr6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.147058 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-config" (OuterVolumeSpecName: "config") pod "87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7" (UID: "87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.152994 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7" (UID: "87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.204649 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdr6s\" (UniqueName: \"kubernetes.io/projected/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-kube-api-access-hdr6s\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.204697 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.204706 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.834407 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" event={"ID":"87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7","Type":"ContainerDied","Data":"858a7d68323cf73642569f5a700d28eee0b02322a63c765b07e0f2c0bb74aefe"} Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.834734 4806 scope.go:117] "RemoveContainer" containerID="6d6f1bb081e4d9609ef5fc110a6d05da448158bf2d1c12676d23807fc8dd6094" Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.834482 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-p2w2s" Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.837397 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b050be-219f-4adc-83b5-52da68b7bade" containerID="2d63dbbfd7bce0dd94b46761eed7fd02d4e8d2cb0c8005147dd820a388ceb9a5" exitCode=0 Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.837472 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ckhq4" event={"ID":"f4b050be-219f-4adc-83b5-52da68b7bade","Type":"ContainerDied","Data":"2d63dbbfd7bce0dd94b46761eed7fd02d4e8d2cb0c8005147dd820a388ceb9a5"} Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.885915 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p2w2s"] Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.887121 4806 scope.go:117] "RemoveContainer" containerID="dfb4fb43d38b4b4a7ccfd0dde865ffa380a9bf9fe33fe75528ba0fe71a12ab64" Nov 27 10:37:23 crc kubenswrapper[4806]: I1127 10:37:23.894308 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p2w2s"] Nov 27 10:37:24 crc kubenswrapper[4806]: I1127 10:37:24.126897 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7" path="/var/lib/kubelet/pods/87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7/volumes" Nov 27 10:37:24 crc kubenswrapper[4806]: I1127 10:37:24.850960 4806 generic.go:334] "Generic (PLEG): container finished" podID="a19e7111-3ef1-45a9-ae31-0d53a2a6615f" containerID="85bec44b35fe89b343437b8e2b5eacf64f5320da66e448ec00dd6effb530b151" exitCode=0 Nov 27 10:37:24 crc kubenswrapper[4806]: I1127 10:37:24.851148 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pc2bp" event={"ID":"a19e7111-3ef1-45a9-ae31-0d53a2a6615f","Type":"ContainerDied","Data":"85bec44b35fe89b343437b8e2b5eacf64f5320da66e448ec00dd6effb530b151"} Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.227897 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-h8mc6"] Nov 27 10:37:25 crc kubenswrapper[4806]: E1127 10:37:25.228411 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3572cba-65ba-47eb-a98a-ff0c4b3feec8" containerName="mariadb-account-create-update" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228443 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3572cba-65ba-47eb-a98a-ff0c4b3feec8" containerName="mariadb-account-create-update" Nov 27 10:37:25 crc kubenswrapper[4806]: E1127 10:37:25.228460 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc617567-5421-4521-8957-4b11667db29e" containerName="mariadb-database-create" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228470 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc617567-5421-4521-8957-4b11667db29e" containerName="mariadb-database-create" Nov 27 10:37:25 crc kubenswrapper[4806]: E1127 10:37:25.228494 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="685cee37-d40c-4ee8-ad87-c67acb12fd60" containerName="mariadb-account-create-update" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228501 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="685cee37-d40c-4ee8-ad87-c67acb12fd60" containerName="mariadb-account-create-update" Nov 27 10:37:25 crc kubenswrapper[4806]: E1127 10:37:25.228514 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b92e4ceb-87fe-4521-b931-9e792f9e4590" containerName="mariadb-database-create" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228522 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="b92e4ceb-87fe-4521-b931-9e792f9e4590" containerName="mariadb-database-create" Nov 27 10:37:25 crc kubenswrapper[4806]: E1127 10:37:25.228534 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5722e8-f491-4a02-836a-f2953334ec7d" containerName="mariadb-database-create" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228541 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5722e8-f491-4a02-836a-f2953334ec7d" containerName="mariadb-database-create" Nov 27 10:37:25 crc kubenswrapper[4806]: E1127 10:37:25.228555 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7" containerName="dnsmasq-dns" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228563 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7" containerName="dnsmasq-dns" Nov 27 10:37:25 crc kubenswrapper[4806]: E1127 10:37:25.228575 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1020f922-5226-4961-aa56-04c97874b6a6" containerName="mariadb-account-create-update" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228583 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="1020f922-5226-4961-aa56-04c97874b6a6" containerName="mariadb-account-create-update" Nov 27 10:37:25 crc kubenswrapper[4806]: E1127 10:37:25.228603 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7" containerName="init" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228610 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7" containerName="init" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228802 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="87df61a9-cd16-4ccb-b8bc-f9fd9cca47d7" containerName="dnsmasq-dns" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228819 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc617567-5421-4521-8957-4b11667db29e" containerName="mariadb-database-create" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228836 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="1020f922-5226-4961-aa56-04c97874b6a6" containerName="mariadb-account-create-update" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228851 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3572cba-65ba-47eb-a98a-ff0c4b3feec8" containerName="mariadb-account-create-update" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228862 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff5722e8-f491-4a02-836a-f2953334ec7d" containerName="mariadb-database-create" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228873 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="685cee37-d40c-4ee8-ad87-c67acb12fd60" containerName="mariadb-account-create-update" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.228886 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="b92e4ceb-87fe-4521-b931-9e792f9e4590" containerName="mariadb-database-create" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.229745 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.241532 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.241542 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-plcm2" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.243009 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-db-sync-config-data\") pod \"glance-db-sync-h8mc6\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.243102 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-combined-ca-bundle\") pod \"glance-db-sync-h8mc6\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.243396 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-config-data\") pod \"glance-db-sync-h8mc6\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.243696 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnctc\" (UniqueName: \"kubernetes.io/projected/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-kube-api-access-nnctc\") pod \"glance-db-sync-h8mc6\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.254348 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-h8mc6"] Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.345858 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-config-data\") pod \"glance-db-sync-h8mc6\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.345937 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnctc\" (UniqueName: \"kubernetes.io/projected/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-kube-api-access-nnctc\") pod \"glance-db-sync-h8mc6\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.345972 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-db-sync-config-data\") pod \"glance-db-sync-h8mc6\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.345992 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-combined-ca-bundle\") pod \"glance-db-sync-h8mc6\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.352700 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-config-data\") pod \"glance-db-sync-h8mc6\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.353482 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-combined-ca-bundle\") pod \"glance-db-sync-h8mc6\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.364098 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-db-sync-config-data\") pod \"glance-db-sync-h8mc6\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.364698 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnctc\" (UniqueName: \"kubernetes.io/projected/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-kube-api-access-nnctc\") pod \"glance-db-sync-h8mc6\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.559805 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.863872 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pc2bp" event={"ID":"a19e7111-3ef1-45a9-ae31-0d53a2a6615f","Type":"ContainerStarted","Data":"0a09695576814fa95ab634c9256a40547e4c8bdc37ed926b7e9e7b61d6a01d2a"} Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.866918 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b050be-219f-4adc-83b5-52da68b7bade" containerID="243f94749037f77765b95cd4a215e368875dddc8a74cf7113cd9da1cfccdcf21" exitCode=0 Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.866954 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ckhq4" event={"ID":"f4b050be-219f-4adc-83b5-52da68b7bade","Type":"ContainerDied","Data":"243f94749037f77765b95cd4a215e368875dddc8a74cf7113cd9da1cfccdcf21"} Nov 27 10:37:25 crc kubenswrapper[4806]: I1127 10:37:25.896778 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pc2bp" podStartSLOduration=2.856134338 podStartE2EDuration="17.89675379s" podCreationTimestamp="2025-11-27 10:37:08 +0000 UTC" firstStartedPulling="2025-11-27 10:37:10.271463254 +0000 UTC m=+934.858054018" lastFinishedPulling="2025-11-27 10:37:25.312082706 +0000 UTC m=+949.898673470" observedRunningTime="2025-11-27 10:37:25.895899797 +0000 UTC m=+950.482490571" watchObservedRunningTime="2025-11-27 10:37:25.89675379 +0000 UTC m=+950.483344554" Nov 27 10:37:26 crc kubenswrapper[4806]: W1127 10:37:26.223909 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aefc973_2bd6_449e_a7e7_afd6a7915c7a.slice/crio-ecb4fbd1fa88d7c7b6953b28a65d3cd1d5b98f5d3c1e77212ff7b4b4c3b39e95 WatchSource:0}: Error finding container ecb4fbd1fa88d7c7b6953b28a65d3cd1d5b98f5d3c1e77212ff7b4b4c3b39e95: Status 404 returned error can't find the container with id ecb4fbd1fa88d7c7b6953b28a65d3cd1d5b98f5d3c1e77212ff7b4b4c3b39e95 Nov 27 10:37:26 crc kubenswrapper[4806]: I1127 10:37:26.246438 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-h8mc6"] Nov 27 10:37:26 crc kubenswrapper[4806]: I1127 10:37:26.878562 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h8mc6" event={"ID":"9aefc973-2bd6-449e-a7e7-afd6a7915c7a","Type":"ContainerStarted","Data":"ecb4fbd1fa88d7c7b6953b28a65d3cd1d5b98f5d3c1e77212ff7b4b4c3b39e95"} Nov 27 10:37:27 crc kubenswrapper[4806]: I1127 10:37:27.899518 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ckhq4" event={"ID":"f4b050be-219f-4adc-83b5-52da68b7bade","Type":"ContainerStarted","Data":"c60b2b72c30c24e1c53e42c9b2fdcf805f177874953274fc1cb6887177bd1620"} Nov 27 10:37:27 crc kubenswrapper[4806]: I1127 10:37:27.923623 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ckhq4" podStartSLOduration=3.818691804 podStartE2EDuration="6.923604376s" podCreationTimestamp="2025-11-27 10:37:21 +0000 UTC" firstStartedPulling="2025-11-27 10:37:23.839897362 +0000 UTC m=+948.426488126" lastFinishedPulling="2025-11-27 10:37:26.944809934 +0000 UTC m=+951.531400698" observedRunningTime="2025-11-27 10:37:27.920936672 +0000 UTC m=+952.507527446" watchObservedRunningTime="2025-11-27 10:37:27.923604376 +0000 UTC m=+952.510195140" Nov 27 10:37:28 crc kubenswrapper[4806]: I1127 10:37:28.183266 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 27 10:37:28 crc kubenswrapper[4806]: I1127 10:37:28.584074 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:28 crc kubenswrapper[4806]: I1127 10:37:28.584549 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:28 crc kubenswrapper[4806]: I1127 10:37:28.657653 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.173596 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.174028 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-f769m" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.356885 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.357960 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.403933 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.440545 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-l82sn-config-jvfv7"] Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.441700 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.453233 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.459663 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-run\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.459856 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-log-ovn\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.459936 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ee90ab60-6e5b-44f5-9445-f874090f7c26-additional-scripts\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.460032 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-run-ovn\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.460113 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlqtp\" (UniqueName: \"kubernetes.io/projected/ee90ab60-6e5b-44f5-9445-f874090f7c26-kube-api-access-zlqtp\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.460196 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee90ab60-6e5b-44f5-9445-f874090f7c26-scripts\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.478929 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l82sn-config-jvfv7"] Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.561747 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlqtp\" (UniqueName: \"kubernetes.io/projected/ee90ab60-6e5b-44f5-9445-f874090f7c26-kube-api-access-zlqtp\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.562177 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee90ab60-6e5b-44f5-9445-f874090f7c26-scripts\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.562558 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-run\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.562697 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-log-ovn\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.562800 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ee90ab60-6e5b-44f5-9445-f874090f7c26-additional-scripts\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.562908 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-run\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.562906 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-log-ovn\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.563109 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-run-ovn\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.563222 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-run-ovn\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.563668 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ee90ab60-6e5b-44f5-9445-f874090f7c26-additional-scripts\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.564475 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee90ab60-6e5b-44f5-9445-f874090f7c26-scripts\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.595238 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlqtp\" (UniqueName: \"kubernetes.io/projected/ee90ab60-6e5b-44f5-9445-f874090f7c26-kube-api-access-zlqtp\") pod \"ovn-controller-l82sn-config-jvfv7\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.681203 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="08137128-e0d1-49d2-abf1-1242c1cb58ce" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.94:5671: connect: connection refused" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.710820 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-l82sn" podUID="2cb892e8-358b-477b-8741-4211a599bf6c" containerName="ovn-controller" probeResult="failure" output=< Nov 27 10:37:31 crc kubenswrapper[4806]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 27 10:37:31 crc kubenswrapper[4806]: > Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.761584 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:31 crc kubenswrapper[4806]: I1127 10:37:31.823395 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="954f26d7-2bdc-45e3-a07b-d9924469f0a1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.95:5671: connect: connection refused" Nov 27 10:37:32 crc kubenswrapper[4806]: I1127 10:37:32.248363 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-l82sn-config-jvfv7"] Nov 27 10:37:33 crc kubenswrapper[4806]: I1127 10:37:33.008169 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:33 crc kubenswrapper[4806]: I1127 10:37:33.066171 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ckhq4"] Nov 27 10:37:34 crc kubenswrapper[4806]: I1127 10:37:34.964609 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ckhq4" podUID="f4b050be-219f-4adc-83b5-52da68b7bade" containerName="registry-server" containerID="cri-o://c60b2b72c30c24e1c53e42c9b2fdcf805f177874953274fc1cb6887177bd1620" gracePeriod=2 Nov 27 10:37:35 crc kubenswrapper[4806]: I1127 10:37:35.976691 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b050be-219f-4adc-83b5-52da68b7bade" containerID="c60b2b72c30c24e1c53e42c9b2fdcf805f177874953274fc1cb6887177bd1620" exitCode=0 Nov 27 10:37:35 crc kubenswrapper[4806]: I1127 10:37:35.976739 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ckhq4" event={"ID":"f4b050be-219f-4adc-83b5-52da68b7bade","Type":"ContainerDied","Data":"c60b2b72c30c24e1c53e42c9b2fdcf805f177874953274fc1cb6887177bd1620"} Nov 27 10:37:36 crc kubenswrapper[4806]: I1127 10:37:36.702615 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-l82sn" podUID="2cb892e8-358b-477b-8741-4211a599bf6c" containerName="ovn-controller" probeResult="failure" output=< Nov 27 10:37:36 crc kubenswrapper[4806]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 27 10:37:36 crc kubenswrapper[4806]: > Nov 27 10:37:38 crc kubenswrapper[4806]: I1127 10:37:38.638086 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pc2bp" Nov 27 10:37:38 crc kubenswrapper[4806]: I1127 10:37:38.717293 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pc2bp"] Nov 27 10:37:38 crc kubenswrapper[4806]: I1127 10:37:38.767742 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jh6qx"] Nov 27 10:37:38 crc kubenswrapper[4806]: I1127 10:37:38.768159 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jh6qx" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" containerName="registry-server" containerID="cri-o://155ad50cfc793da40086147207ae8c331a2365531f78f50833a16d8ac3e7800e" gracePeriod=2 Nov 27 10:37:39 crc kubenswrapper[4806]: I1127 10:37:39.042407 4806 generic.go:334] "Generic (PLEG): container finished" podID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" containerID="155ad50cfc793da40086147207ae8c331a2365531f78f50833a16d8ac3e7800e" exitCode=0 Nov 27 10:37:39 crc kubenswrapper[4806]: I1127 10:37:39.042764 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh6qx" event={"ID":"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d","Type":"ContainerDied","Data":"155ad50cfc793da40086147207ae8c331a2365531f78f50833a16d8ac3e7800e"} Nov 27 10:37:40 crc kubenswrapper[4806]: I1127 10:37:40.879744 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.039790 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.065165 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hflm\" (UniqueName: \"kubernetes.io/projected/f4b050be-219f-4adc-83b5-52da68b7bade-kube-api-access-7hflm\") pod \"f4b050be-219f-4adc-83b5-52da68b7bade\" (UID: \"f4b050be-219f-4adc-83b5-52da68b7bade\") " Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.065661 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4b050be-219f-4adc-83b5-52da68b7bade-utilities\") pod \"f4b050be-219f-4adc-83b5-52da68b7bade\" (UID: \"f4b050be-219f-4adc-83b5-52da68b7bade\") " Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.065891 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4b050be-219f-4adc-83b5-52da68b7bade-catalog-content\") pod \"f4b050be-219f-4adc-83b5-52da68b7bade\" (UID: \"f4b050be-219f-4adc-83b5-52da68b7bade\") " Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.067841 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4b050be-219f-4adc-83b5-52da68b7bade-utilities" (OuterVolumeSpecName: "utilities") pod "f4b050be-219f-4adc-83b5-52da68b7bade" (UID: "f4b050be-219f-4adc-83b5-52da68b7bade"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.076306 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4b050be-219f-4adc-83b5-52da68b7bade-kube-api-access-7hflm" (OuterVolumeSpecName: "kube-api-access-7hflm") pod "f4b050be-219f-4adc-83b5-52da68b7bade" (UID: "f4b050be-219f-4adc-83b5-52da68b7bade"). InnerVolumeSpecName "kube-api-access-7hflm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.108490 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jh6qx" event={"ID":"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d","Type":"ContainerDied","Data":"5e04dea91ab4046f41d4659c58eea636d8b65efafa5300dbb7de84ae1a836597"} Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.109260 4806 scope.go:117] "RemoveContainer" containerID="155ad50cfc793da40086147207ae8c331a2365531f78f50833a16d8ac3e7800e" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.108751 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jh6qx" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.153044 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l82sn-config-jvfv7" event={"ID":"ee90ab60-6e5b-44f5-9445-f874090f7c26","Type":"ContainerStarted","Data":"a30671900ed67563a7771f57e42fb02e9c97d36bfd380b80aad0bcd6d38ca86e"} Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.153105 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l82sn-config-jvfv7" event={"ID":"ee90ab60-6e5b-44f5-9445-f874090f7c26","Type":"ContainerStarted","Data":"ac180e905b0cb93bb98745203b88f11441b5432c6705a4adc8d62d09d63135c6"} Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.158752 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4b050be-219f-4adc-83b5-52da68b7bade-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4b050be-219f-4adc-83b5-52da68b7bade" (UID: "f4b050be-219f-4adc-83b5-52da68b7bade"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.169644 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-utilities\") pod \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\" (UID: \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\") " Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.169755 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-catalog-content\") pod \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\" (UID: \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\") " Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.169884 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct9jr\" (UniqueName: \"kubernetes.io/projected/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-kube-api-access-ct9jr\") pod \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\" (UID: \"a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d\") " Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.171139 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-utilities" (OuterVolumeSpecName: "utilities") pod "a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" (UID: "a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.174079 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4b050be-219f-4adc-83b5-52da68b7bade-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.174899 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hflm\" (UniqueName: \"kubernetes.io/projected/f4b050be-219f-4adc-83b5-52da68b7bade-kube-api-access-7hflm\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.174924 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4b050be-219f-4adc-83b5-52da68b7bade-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.184066 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-l82sn-config-jvfv7" podStartSLOduration=10.184045204 podStartE2EDuration="10.184045204s" podCreationTimestamp="2025-11-27 10:37:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:37:41.181619226 +0000 UTC m=+965.768209990" watchObservedRunningTime="2025-11-27 10:37:41.184045204 +0000 UTC m=+965.770635968" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.228811 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-kube-api-access-ct9jr" (OuterVolumeSpecName: "kube-api-access-ct9jr") pod "a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" (UID: "a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d"). InnerVolumeSpecName "kube-api-access-ct9jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.240652 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ckhq4" event={"ID":"f4b050be-219f-4adc-83b5-52da68b7bade","Type":"ContainerDied","Data":"43950de7e7bc7fbaf8555ce23c6fe60f2fd40facf06f39e1176902affe651726"} Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.240717 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ckhq4" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.283665 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" (UID: "a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.294447 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ckhq4"] Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.294490 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.294523 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.294537 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct9jr\" (UniqueName: \"kubernetes.io/projected/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d-kube-api-access-ct9jr\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.303103 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ckhq4"] Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.317254 4806 scope.go:117] "RemoveContainer" containerID="f295f1e491c669bef133f65ff7de5819cae5c506e39d7507a6f15d5b0759d42a" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.370784 4806 scope.go:117] "RemoveContainer" containerID="c249c10cd2c23aa758eee36e2a9ef2ca465977c37fe93b2d3aa0cf104f924030" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.393253 4806 scope.go:117] "RemoveContainer" containerID="c60b2b72c30c24e1c53e42c9b2fdcf805f177874953274fc1cb6887177bd1620" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.493317 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jh6qx"] Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.508187 4806 scope.go:117] "RemoveContainer" containerID="243f94749037f77765b95cd4a215e368875dddc8a74cf7113cd9da1cfccdcf21" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.511151 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jh6qx"] Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.570473 4806 scope.go:117] "RemoveContainer" containerID="2d63dbbfd7bce0dd94b46761eed7fd02d4e8d2cb0c8005147dd820a388ceb9a5" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.681484 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.823211 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:37:41 crc kubenswrapper[4806]: I1127 10:37:41.921156 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-l82sn" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.128261 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" path="/var/lib/kubelet/pods/a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d/volumes" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.129177 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b050be-219f-4adc-83b5-52da68b7bade" path="/var/lib/kubelet/pods/f4b050be-219f-4adc-83b5-52da68b7bade/volumes" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.292193 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h8mc6" event={"ID":"9aefc973-2bd6-449e-a7e7-afd6a7915c7a","Type":"ContainerStarted","Data":"856d1cbfd076b861d6ec0311109278411b8f03ca997a7f6e4b585020b52cb83f"} Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.450766 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-h8mc6" podStartSLOduration=3.054322871 podStartE2EDuration="17.45073474s" podCreationTimestamp="2025-11-27 10:37:25 +0000 UTC" firstStartedPulling="2025-11-27 10:37:26.226108703 +0000 UTC m=+950.812699467" lastFinishedPulling="2025-11-27 10:37:40.622520572 +0000 UTC m=+965.209111336" observedRunningTime="2025-11-27 10:37:42.368015447 +0000 UTC m=+966.954606221" watchObservedRunningTime="2025-11-27 10:37:42.45073474 +0000 UTC m=+967.037325504" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.459920 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-f488s"] Nov 27 10:37:42 crc kubenswrapper[4806]: E1127 10:37:42.461163 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b050be-219f-4adc-83b5-52da68b7bade" containerName="extract-utilities" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.461218 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b050be-219f-4adc-83b5-52da68b7bade" containerName="extract-utilities" Nov 27 10:37:42 crc kubenswrapper[4806]: E1127 10:37:42.467083 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b050be-219f-4adc-83b5-52da68b7bade" containerName="registry-server" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.467155 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b050be-219f-4adc-83b5-52da68b7bade" containerName="registry-server" Nov 27 10:37:42 crc kubenswrapper[4806]: E1127 10:37:42.467188 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" containerName="extract-content" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.467204 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" containerName="extract-content" Nov 27 10:37:42 crc kubenswrapper[4806]: E1127 10:37:42.467323 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b050be-219f-4adc-83b5-52da68b7bade" containerName="extract-content" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.467345 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b050be-219f-4adc-83b5-52da68b7bade" containerName="extract-content" Nov 27 10:37:42 crc kubenswrapper[4806]: E1127 10:37:42.467370 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" containerName="extract-utilities" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.467381 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" containerName="extract-utilities" Nov 27 10:37:42 crc kubenswrapper[4806]: E1127 10:37:42.467399 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" containerName="registry-server" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.467405 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" containerName="registry-server" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.471712 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b050be-219f-4adc-83b5-52da68b7bade" containerName="registry-server" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.471774 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7c21192-8990-4c0d-bd2a-e3ba5c1ba11d" containerName="registry-server" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.472730 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-f488s" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.505371 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-f488s"] Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.591989 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-3071-account-create-update-2g8jt"] Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.593075 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3071-account-create-update-2g8jt" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.609789 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.620498 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3071-account-create-update-2g8jt"] Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.627572 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsjqs\" (UniqueName: \"kubernetes.io/projected/a9270118-f53f-4b38-98e6-03469db9302d-kube-api-access-fsjqs\") pod \"barbican-db-create-f488s\" (UID: \"a9270118-f53f-4b38-98e6-03469db9302d\") " pod="openstack/barbican-db-create-f488s" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.627784 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9270118-f53f-4b38-98e6-03469db9302d-operator-scripts\") pod \"barbican-db-create-f488s\" (UID: \"a9270118-f53f-4b38-98e6-03469db9302d\") " pod="openstack/barbican-db-create-f488s" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.670855 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-rhbrc"] Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.672124 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rhbrc" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.729841 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9270118-f53f-4b38-98e6-03469db9302d-operator-scripts\") pod \"barbican-db-create-f488s\" (UID: \"a9270118-f53f-4b38-98e6-03469db9302d\") " pod="openstack/barbican-db-create-f488s" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.729925 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-242vj\" (UniqueName: \"kubernetes.io/projected/a3e71c8a-a010-49f6-9e79-c868ec3243d9-kube-api-access-242vj\") pod \"cinder-db-create-rhbrc\" (UID: \"a3e71c8a-a010-49f6-9e79-c868ec3243d9\") " pod="openstack/cinder-db-create-rhbrc" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.730034 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df636244-5318-44a8-857b-2c68af37b10a-operator-scripts\") pod \"barbican-3071-account-create-update-2g8jt\" (UID: \"df636244-5318-44a8-857b-2c68af37b10a\") " pod="openstack/barbican-3071-account-create-update-2g8jt" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.730077 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsjqs\" (UniqueName: \"kubernetes.io/projected/a9270118-f53f-4b38-98e6-03469db9302d-kube-api-access-fsjqs\") pod \"barbican-db-create-f488s\" (UID: \"a9270118-f53f-4b38-98e6-03469db9302d\") " pod="openstack/barbican-db-create-f488s" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.730117 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3e71c8a-a010-49f6-9e79-c868ec3243d9-operator-scripts\") pod \"cinder-db-create-rhbrc\" (UID: \"a3e71c8a-a010-49f6-9e79-c868ec3243d9\") " pod="openstack/cinder-db-create-rhbrc" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.730159 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2nlk\" (UniqueName: \"kubernetes.io/projected/df636244-5318-44a8-857b-2c68af37b10a-kube-api-access-h2nlk\") pod \"barbican-3071-account-create-update-2g8jt\" (UID: \"df636244-5318-44a8-857b-2c68af37b10a\") " pod="openstack/barbican-3071-account-create-update-2g8jt" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.730847 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9270118-f53f-4b38-98e6-03469db9302d-operator-scripts\") pod \"barbican-db-create-f488s\" (UID: \"a9270118-f53f-4b38-98e6-03469db9302d\") " pod="openstack/barbican-db-create-f488s" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.746786 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rhbrc"] Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.779094 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsjqs\" (UniqueName: \"kubernetes.io/projected/a9270118-f53f-4b38-98e6-03469db9302d-kube-api-access-fsjqs\") pod \"barbican-db-create-f488s\" (UID: \"a9270118-f53f-4b38-98e6-03469db9302d\") " pod="openstack/barbican-db-create-f488s" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.799699 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-f488s" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.841422 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df636244-5318-44a8-857b-2c68af37b10a-operator-scripts\") pod \"barbican-3071-account-create-update-2g8jt\" (UID: \"df636244-5318-44a8-857b-2c68af37b10a\") " pod="openstack/barbican-3071-account-create-update-2g8jt" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.841494 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3e71c8a-a010-49f6-9e79-c868ec3243d9-operator-scripts\") pod \"cinder-db-create-rhbrc\" (UID: \"a3e71c8a-a010-49f6-9e79-c868ec3243d9\") " pod="openstack/cinder-db-create-rhbrc" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.841528 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2nlk\" (UniqueName: \"kubernetes.io/projected/df636244-5318-44a8-857b-2c68af37b10a-kube-api-access-h2nlk\") pod \"barbican-3071-account-create-update-2g8jt\" (UID: \"df636244-5318-44a8-857b-2c68af37b10a\") " pod="openstack/barbican-3071-account-create-update-2g8jt" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.841589 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-242vj\" (UniqueName: \"kubernetes.io/projected/a3e71c8a-a010-49f6-9e79-c868ec3243d9-kube-api-access-242vj\") pod \"cinder-db-create-rhbrc\" (UID: \"a3e71c8a-a010-49f6-9e79-c868ec3243d9\") " pod="openstack/cinder-db-create-rhbrc" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.842852 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df636244-5318-44a8-857b-2c68af37b10a-operator-scripts\") pod \"barbican-3071-account-create-update-2g8jt\" (UID: \"df636244-5318-44a8-857b-2c68af37b10a\") " pod="openstack/barbican-3071-account-create-update-2g8jt" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.842897 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3e71c8a-a010-49f6-9e79-c868ec3243d9-operator-scripts\") pod \"cinder-db-create-rhbrc\" (UID: \"a3e71c8a-a010-49f6-9e79-c868ec3243d9\") " pod="openstack/cinder-db-create-rhbrc" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.894574 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2nlk\" (UniqueName: \"kubernetes.io/projected/df636244-5318-44a8-857b-2c68af37b10a-kube-api-access-h2nlk\") pod \"barbican-3071-account-create-update-2g8jt\" (UID: \"df636244-5318-44a8-857b-2c68af37b10a\") " pod="openstack/barbican-3071-account-create-update-2g8jt" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.905058 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-242vj\" (UniqueName: \"kubernetes.io/projected/a3e71c8a-a010-49f6-9e79-c868ec3243d9-kube-api-access-242vj\") pod \"cinder-db-create-rhbrc\" (UID: \"a3e71c8a-a010-49f6-9e79-c868ec3243d9\") " pod="openstack/cinder-db-create-rhbrc" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.907967 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-1c99-account-create-update-6mm4j"] Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.909182 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1c99-account-create-update-6mm4j" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.913890 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3071-account-create-update-2g8jt" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.936434 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.960232 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-1c99-account-create-update-6mm4j"] Nov 27 10:37:42 crc kubenswrapper[4806]: I1127 10:37:42.986797 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rhbrc" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.048661 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5mxz\" (UniqueName: \"kubernetes.io/projected/82baaafd-2458-40d0-b859-8171f9a6db0d-kube-api-access-p5mxz\") pod \"cinder-1c99-account-create-update-6mm4j\" (UID: \"82baaafd-2458-40d0-b859-8171f9a6db0d\") " pod="openstack/cinder-1c99-account-create-update-6mm4j" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.048763 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82baaafd-2458-40d0-b859-8171f9a6db0d-operator-scripts\") pod \"cinder-1c99-account-create-update-6mm4j\" (UID: \"82baaafd-2458-40d0-b859-8171f9a6db0d\") " pod="openstack/cinder-1c99-account-create-update-6mm4j" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.107293 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-xptpp"] Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.109217 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xptpp" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.148710 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xptpp"] Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.151100 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82baaafd-2458-40d0-b859-8171f9a6db0d-operator-scripts\") pod \"cinder-1c99-account-create-update-6mm4j\" (UID: \"82baaafd-2458-40d0-b859-8171f9a6db0d\") " pod="openstack/cinder-1c99-account-create-update-6mm4j" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.151299 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5mxz\" (UniqueName: \"kubernetes.io/projected/82baaafd-2458-40d0-b859-8171f9a6db0d-kube-api-access-p5mxz\") pod \"cinder-1c99-account-create-update-6mm4j\" (UID: \"82baaafd-2458-40d0-b859-8171f9a6db0d\") " pod="openstack/cinder-1c99-account-create-update-6mm4j" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.152720 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82baaafd-2458-40d0-b859-8171f9a6db0d-operator-scripts\") pod \"cinder-1c99-account-create-update-6mm4j\" (UID: \"82baaafd-2458-40d0-b859-8171f9a6db0d\") " pod="openstack/cinder-1c99-account-create-update-6mm4j" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.194545 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5mxz\" (UniqueName: \"kubernetes.io/projected/82baaafd-2458-40d0-b859-8171f9a6db0d-kube-api-access-p5mxz\") pod \"cinder-1c99-account-create-update-6mm4j\" (UID: \"82baaafd-2458-40d0-b859-8171f9a6db0d\") " pod="openstack/cinder-1c99-account-create-update-6mm4j" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.207146 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-3a87-account-create-update-8zm2q"] Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.208578 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3a87-account-create-update-8zm2q" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.212642 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.246635 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3a87-account-create-update-8zm2q"] Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.262127 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2drn\" (UniqueName: \"kubernetes.io/projected/f7143fe2-d8ae-4d55-a11c-edb06cf922f7-kube-api-access-m2drn\") pod \"neutron-db-create-xptpp\" (UID: \"f7143fe2-d8ae-4d55-a11c-edb06cf922f7\") " pod="openstack/neutron-db-create-xptpp" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.262801 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7143fe2-d8ae-4d55-a11c-edb06cf922f7-operator-scripts\") pod \"neutron-db-create-xptpp\" (UID: \"f7143fe2-d8ae-4d55-a11c-edb06cf922f7\") " pod="openstack/neutron-db-create-xptpp" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.308105 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1c99-account-create-update-6mm4j" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.374871 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7143fe2-d8ae-4d55-a11c-edb06cf922f7-operator-scripts\") pod \"neutron-db-create-xptpp\" (UID: \"f7143fe2-d8ae-4d55-a11c-edb06cf922f7\") " pod="openstack/neutron-db-create-xptpp" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.375657 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb5dr\" (UniqueName: \"kubernetes.io/projected/679580af-7a44-485f-8b79-f79b702d85eb-kube-api-access-nb5dr\") pod \"neutron-3a87-account-create-update-8zm2q\" (UID: \"679580af-7a44-485f-8b79-f79b702d85eb\") " pod="openstack/neutron-3a87-account-create-update-8zm2q" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.376719 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2drn\" (UniqueName: \"kubernetes.io/projected/f7143fe2-d8ae-4d55-a11c-edb06cf922f7-kube-api-access-m2drn\") pod \"neutron-db-create-xptpp\" (UID: \"f7143fe2-d8ae-4d55-a11c-edb06cf922f7\") " pod="openstack/neutron-db-create-xptpp" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.379086 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/679580af-7a44-485f-8b79-f79b702d85eb-operator-scripts\") pod \"neutron-3a87-account-create-update-8zm2q\" (UID: \"679580af-7a44-485f-8b79-f79b702d85eb\") " pod="openstack/neutron-3a87-account-create-update-8zm2q" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.392710 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7143fe2-d8ae-4d55-a11c-edb06cf922f7-operator-scripts\") pod \"neutron-db-create-xptpp\" (UID: \"f7143fe2-d8ae-4d55-a11c-edb06cf922f7\") " pod="openstack/neutron-db-create-xptpp" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.392797 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-6vcz5"] Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.394540 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6vcz5" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.401212 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.403417 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.403618 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.409877 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-6vcz5"] Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.411560 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxgbr" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.435712 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2drn\" (UniqueName: \"kubernetes.io/projected/f7143fe2-d8ae-4d55-a11c-edb06cf922f7-kube-api-access-m2drn\") pod \"neutron-db-create-xptpp\" (UID: \"f7143fe2-d8ae-4d55-a11c-edb06cf922f7\") " pod="openstack/neutron-db-create-xptpp" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.463658 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xptpp" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.482463 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb5dr\" (UniqueName: \"kubernetes.io/projected/679580af-7a44-485f-8b79-f79b702d85eb-kube-api-access-nb5dr\") pod \"neutron-3a87-account-create-update-8zm2q\" (UID: \"679580af-7a44-485f-8b79-f79b702d85eb\") " pod="openstack/neutron-3a87-account-create-update-8zm2q" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.482559 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg2xq\" (UniqueName: \"kubernetes.io/projected/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-kube-api-access-qg2xq\") pod \"keystone-db-sync-6vcz5\" (UID: \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\") " pod="openstack/keystone-db-sync-6vcz5" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.482619 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-config-data\") pod \"keystone-db-sync-6vcz5\" (UID: \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\") " pod="openstack/keystone-db-sync-6vcz5" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.482671 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/679580af-7a44-485f-8b79-f79b702d85eb-operator-scripts\") pod \"neutron-3a87-account-create-update-8zm2q\" (UID: \"679580af-7a44-485f-8b79-f79b702d85eb\") " pod="openstack/neutron-3a87-account-create-update-8zm2q" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.482726 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-combined-ca-bundle\") pod \"keystone-db-sync-6vcz5\" (UID: \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\") " pod="openstack/keystone-db-sync-6vcz5" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.485416 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/679580af-7a44-485f-8b79-f79b702d85eb-operator-scripts\") pod \"neutron-3a87-account-create-update-8zm2q\" (UID: \"679580af-7a44-485f-8b79-f79b702d85eb\") " pod="openstack/neutron-3a87-account-create-update-8zm2q" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.528982 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb5dr\" (UniqueName: \"kubernetes.io/projected/679580af-7a44-485f-8b79-f79b702d85eb-kube-api-access-nb5dr\") pod \"neutron-3a87-account-create-update-8zm2q\" (UID: \"679580af-7a44-485f-8b79-f79b702d85eb\") " pod="openstack/neutron-3a87-account-create-update-8zm2q" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.573085 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3a87-account-create-update-8zm2q" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.587040 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-config-data\") pod \"keystone-db-sync-6vcz5\" (UID: \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\") " pod="openstack/keystone-db-sync-6vcz5" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.587136 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-combined-ca-bundle\") pod \"keystone-db-sync-6vcz5\" (UID: \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\") " pod="openstack/keystone-db-sync-6vcz5" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.587217 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg2xq\" (UniqueName: \"kubernetes.io/projected/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-kube-api-access-qg2xq\") pod \"keystone-db-sync-6vcz5\" (UID: \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\") " pod="openstack/keystone-db-sync-6vcz5" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.592255 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-config-data\") pod \"keystone-db-sync-6vcz5\" (UID: \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\") " pod="openstack/keystone-db-sync-6vcz5" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.600993 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-combined-ca-bundle\") pod \"keystone-db-sync-6vcz5\" (UID: \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\") " pod="openstack/keystone-db-sync-6vcz5" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.614314 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg2xq\" (UniqueName: \"kubernetes.io/projected/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-kube-api-access-qg2xq\") pod \"keystone-db-sync-6vcz5\" (UID: \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\") " pod="openstack/keystone-db-sync-6vcz5" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.735786 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6vcz5" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.846959 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="b30ea2f1-af76-4a43-842c-577ff4f35a20" containerName="galera" probeResult="failure" output="command timed out" Nov 27 10:37:43 crc kubenswrapper[4806]: I1127 10:37:43.926094 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3071-account-create-update-2g8jt"] Nov 27 10:37:44 crc kubenswrapper[4806]: I1127 10:37:44.205371 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-f488s"] Nov 27 10:37:44 crc kubenswrapper[4806]: I1127 10:37:44.454288 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3071-account-create-update-2g8jt" event={"ID":"df636244-5318-44a8-857b-2c68af37b10a","Type":"ContainerStarted","Data":"9dbefdb03453c954efa5edbfd746ee3de894e1c7c5e2de635eac0cb381a5b8f3"} Nov 27 10:37:44 crc kubenswrapper[4806]: I1127 10:37:44.464820 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-f488s" event={"ID":"a9270118-f53f-4b38-98e6-03469db9302d","Type":"ContainerStarted","Data":"d1a458be97a6aa38586471d651b06e63063a5c6a95ff77ea2b5e1e071099d9fd"} Nov 27 10:37:44 crc kubenswrapper[4806]: I1127 10:37:44.472284 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rhbrc"] Nov 27 10:37:44 crc kubenswrapper[4806]: I1127 10:37:44.478826 4806 generic.go:334] "Generic (PLEG): container finished" podID="ee90ab60-6e5b-44f5-9445-f874090f7c26" containerID="a30671900ed67563a7771f57e42fb02e9c97d36bfd380b80aad0bcd6d38ca86e" exitCode=0 Nov 27 10:37:44 crc kubenswrapper[4806]: I1127 10:37:44.478879 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l82sn-config-jvfv7" event={"ID":"ee90ab60-6e5b-44f5-9445-f874090f7c26","Type":"ContainerDied","Data":"a30671900ed67563a7771f57e42fb02e9c97d36bfd380b80aad0bcd6d38ca86e"} Nov 27 10:37:44 crc kubenswrapper[4806]: I1127 10:37:44.673541 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-1c99-account-create-update-6mm4j"] Nov 27 10:37:44 crc kubenswrapper[4806]: I1127 10:37:44.721084 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xptpp"] Nov 27 10:37:44 crc kubenswrapper[4806]: I1127 10:37:44.872564 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3a87-account-create-update-8zm2q"] Nov 27 10:37:44 crc kubenswrapper[4806]: I1127 10:37:44.921179 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-6vcz5"] Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.489459 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3a87-account-create-update-8zm2q" event={"ID":"679580af-7a44-485f-8b79-f79b702d85eb","Type":"ContainerStarted","Data":"85afbb8a4561fd7353250eb568d0df936c03dc313d30b9e1f4ce22eaaa1ae12e"} Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.491247 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3a87-account-create-update-8zm2q" event={"ID":"679580af-7a44-485f-8b79-f79b702d85eb","Type":"ContainerStarted","Data":"1a407f5950a84047e92f9112ab582c546a8061e82ed0b79216d7bf8231edccb4"} Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.497818 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xptpp" event={"ID":"f7143fe2-d8ae-4d55-a11c-edb06cf922f7","Type":"ContainerStarted","Data":"f6b40b2335c910e068c63dab1fac7deb316f1213e1f22109ff58b5a9dabe6b03"} Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.497893 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xptpp" event={"ID":"f7143fe2-d8ae-4d55-a11c-edb06cf922f7","Type":"ContainerStarted","Data":"af94e764c9a902b653eb47642fffd8ce8208bf699bb6b2690f428d5e6be2aa79"} Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.505261 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6vcz5" event={"ID":"c7484e21-2c80-49d6-a9ec-6489bc40ec4f","Type":"ContainerStarted","Data":"d286b5f121341c8d7228c373cd17401952e4de446db8a52b75b35a2d5735fb8d"} Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.535770 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rhbrc" event={"ID":"a3e71c8a-a010-49f6-9e79-c868ec3243d9","Type":"ContainerStarted","Data":"b4a0eb7d0744a08010adc745dcfacc358029c1457f0b148d283dba18fe7bcd90"} Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.536084 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rhbrc" event={"ID":"a3e71c8a-a010-49f6-9e79-c868ec3243d9","Type":"ContainerStarted","Data":"d23551eb8426ee5dbefe6bc762d172b8c2ecc3d5fd8aeceed84cc17ae6ef6f88"} Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.547839 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3071-account-create-update-2g8jt" event={"ID":"df636244-5318-44a8-857b-2c68af37b10a","Type":"ContainerStarted","Data":"eb8d3083c3fa65165665e80e3941c7c11b2e95bac5b598a421fbb172b62015b2"} Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.559922 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-f488s" event={"ID":"a9270118-f53f-4b38-98e6-03469db9302d","Type":"ContainerStarted","Data":"550e722b7e0b4143095a89bd42da2fc92a7788d6fcc069823e4b4bd9e326ee21"} Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.569515 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1c99-account-create-update-6mm4j" event={"ID":"82baaafd-2458-40d0-b859-8171f9a6db0d","Type":"ContainerStarted","Data":"af0b08e34b17f0a2422e31bf51a54bb63701250b03bfa4fce101bfa1b663488d"} Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.675201 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-rhbrc" podStartSLOduration=3.675174388 podStartE2EDuration="3.675174388s" podCreationTimestamp="2025-11-27 10:37:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:37:45.65867043 +0000 UTC m=+970.245261194" watchObservedRunningTime="2025-11-27 10:37:45.675174388 +0000 UTC m=+970.261765152" Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.677928 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-3a87-account-create-update-8zm2q" podStartSLOduration=2.677921994 podStartE2EDuration="2.677921994s" podCreationTimestamp="2025-11-27 10:37:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:37:45.589515212 +0000 UTC m=+970.176105976" watchObservedRunningTime="2025-11-27 10:37:45.677921994 +0000 UTC m=+970.264512758" Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.693659 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-f488s" podStartSLOduration=3.69363897 podStartE2EDuration="3.69363897s" podCreationTimestamp="2025-11-27 10:37:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:37:45.690811412 +0000 UTC m=+970.277402176" watchObservedRunningTime="2025-11-27 10:37:45.69363897 +0000 UTC m=+970.280229734" Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.754156 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-3071-account-create-update-2g8jt" podStartSLOduration=3.754136167 podStartE2EDuration="3.754136167s" podCreationTimestamp="2025-11-27 10:37:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:37:45.752553654 +0000 UTC m=+970.339144418" watchObservedRunningTime="2025-11-27 10:37:45.754136167 +0000 UTC m=+970.340726931" Nov 27 10:37:45 crc kubenswrapper[4806]: I1127 10:37:45.824819 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-xptpp" podStartSLOduration=2.8247994370000002 podStartE2EDuration="2.824799437s" podCreationTimestamp="2025-11-27 10:37:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:37:45.81263737 +0000 UTC m=+970.399228134" watchObservedRunningTime="2025-11-27 10:37:45.824799437 +0000 UTC m=+970.411390191" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.351692 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.520426 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ee90ab60-6e5b-44f5-9445-f874090f7c26-additional-scripts\") pod \"ee90ab60-6e5b-44f5-9445-f874090f7c26\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.520616 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee90ab60-6e5b-44f5-9445-f874090f7c26-scripts\") pod \"ee90ab60-6e5b-44f5-9445-f874090f7c26\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.520695 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-log-ovn\") pod \"ee90ab60-6e5b-44f5-9445-f874090f7c26\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.520893 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlqtp\" (UniqueName: \"kubernetes.io/projected/ee90ab60-6e5b-44f5-9445-f874090f7c26-kube-api-access-zlqtp\") pod \"ee90ab60-6e5b-44f5-9445-f874090f7c26\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.520916 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-run-ovn\") pod \"ee90ab60-6e5b-44f5-9445-f874090f7c26\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.520998 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-run\") pod \"ee90ab60-6e5b-44f5-9445-f874090f7c26\" (UID: \"ee90ab60-6e5b-44f5-9445-f874090f7c26\") " Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.521540 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-run" (OuterVolumeSpecName: "var-run") pod "ee90ab60-6e5b-44f5-9445-f874090f7c26" (UID: "ee90ab60-6e5b-44f5-9445-f874090f7c26"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.522063 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "ee90ab60-6e5b-44f5-9445-f874090f7c26" (UID: "ee90ab60-6e5b-44f5-9445-f874090f7c26"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.522219 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "ee90ab60-6e5b-44f5-9445-f874090f7c26" (UID: "ee90ab60-6e5b-44f5-9445-f874090f7c26"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.522543 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee90ab60-6e5b-44f5-9445-f874090f7c26-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "ee90ab60-6e5b-44f5-9445-f874090f7c26" (UID: "ee90ab60-6e5b-44f5-9445-f874090f7c26"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.525250 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee90ab60-6e5b-44f5-9445-f874090f7c26-scripts" (OuterVolumeSpecName: "scripts") pod "ee90ab60-6e5b-44f5-9445-f874090f7c26" (UID: "ee90ab60-6e5b-44f5-9445-f874090f7c26"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.539426 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee90ab60-6e5b-44f5-9445-f874090f7c26-kube-api-access-zlqtp" (OuterVolumeSpecName: "kube-api-access-zlqtp") pod "ee90ab60-6e5b-44f5-9445-f874090f7c26" (UID: "ee90ab60-6e5b-44f5-9445-f874090f7c26"). InnerVolumeSpecName "kube-api-access-zlqtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.584886 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-l82sn-config-jvfv7" event={"ID":"ee90ab60-6e5b-44f5-9445-f874090f7c26","Type":"ContainerDied","Data":"ac180e905b0cb93bb98745203b88f11441b5432c6705a4adc8d62d09d63135c6"} Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.585259 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac180e905b0cb93bb98745203b88f11441b5432c6705a4adc8d62d09d63135c6" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.585166 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-l82sn-config-jvfv7" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.593135 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1c99-account-create-update-6mm4j" event={"ID":"82baaafd-2458-40d0-b859-8171f9a6db0d","Type":"ContainerStarted","Data":"4f96c0b3875c99592487a94e9960ec45541a81e5c5dcb5a45a909916cd0e015a"} Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.624551 4806 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-run\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.625403 4806 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ee90ab60-6e5b-44f5-9445-f874090f7c26-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.625510 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee90ab60-6e5b-44f5-9445-f874090f7c26-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.634330 4806 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.634382 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlqtp\" (UniqueName: \"kubernetes.io/projected/ee90ab60-6e5b-44f5-9445-f874090f7c26-kube-api-access-zlqtp\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.634395 4806 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ee90ab60-6e5b-44f5-9445-f874090f7c26-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.708739 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-1c99-account-create-update-6mm4j" podStartSLOduration=4.708716549 podStartE2EDuration="4.708716549s" podCreationTimestamp="2025-11-27 10:37:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:37:46.665018047 +0000 UTC m=+971.251608811" watchObservedRunningTime="2025-11-27 10:37:46.708716549 +0000 UTC m=+971.295307313" Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.784688 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-l82sn-config-jvfv7"] Nov 27 10:37:46 crc kubenswrapper[4806]: I1127 10:37:46.803754 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-l82sn-config-jvfv7"] Nov 27 10:37:47 crc kubenswrapper[4806]: I1127 10:37:47.604985 4806 generic.go:334] "Generic (PLEG): container finished" podID="a9270118-f53f-4b38-98e6-03469db9302d" containerID="550e722b7e0b4143095a89bd42da2fc92a7788d6fcc069823e4b4bd9e326ee21" exitCode=0 Nov 27 10:37:47 crc kubenswrapper[4806]: I1127 10:37:47.605989 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-f488s" event={"ID":"a9270118-f53f-4b38-98e6-03469db9302d","Type":"ContainerDied","Data":"550e722b7e0b4143095a89bd42da2fc92a7788d6fcc069823e4b4bd9e326ee21"} Nov 27 10:37:48 crc kubenswrapper[4806]: I1127 10:37:48.130479 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee90ab60-6e5b-44f5-9445-f874090f7c26" path="/var/lib/kubelet/pods/ee90ab60-6e5b-44f5-9445-f874090f7c26/volumes" Nov 27 10:37:48 crc kubenswrapper[4806]: I1127 10:37:48.615780 4806 generic.go:334] "Generic (PLEG): container finished" podID="df636244-5318-44a8-857b-2c68af37b10a" containerID="eb8d3083c3fa65165665e80e3941c7c11b2e95bac5b598a421fbb172b62015b2" exitCode=0 Nov 27 10:37:48 crc kubenswrapper[4806]: I1127 10:37:48.615850 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3071-account-create-update-2g8jt" event={"ID":"df636244-5318-44a8-857b-2c68af37b10a","Type":"ContainerDied","Data":"eb8d3083c3fa65165665e80e3941c7c11b2e95bac5b598a421fbb172b62015b2"} Nov 27 10:37:48 crc kubenswrapper[4806]: I1127 10:37:48.632548 4806 generic.go:334] "Generic (PLEG): container finished" podID="82baaafd-2458-40d0-b859-8171f9a6db0d" containerID="4f96c0b3875c99592487a94e9960ec45541a81e5c5dcb5a45a909916cd0e015a" exitCode=0 Nov 27 10:37:48 crc kubenswrapper[4806]: I1127 10:37:48.632655 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1c99-account-create-update-6mm4j" event={"ID":"82baaafd-2458-40d0-b859-8171f9a6db0d","Type":"ContainerDied","Data":"4f96c0b3875c99592487a94e9960ec45541a81e5c5dcb5a45a909916cd0e015a"} Nov 27 10:37:48 crc kubenswrapper[4806]: I1127 10:37:48.638037 4806 generic.go:334] "Generic (PLEG): container finished" podID="a3e71c8a-a010-49f6-9e79-c868ec3243d9" containerID="b4a0eb7d0744a08010adc745dcfacc358029c1457f0b148d283dba18fe7bcd90" exitCode=0 Nov 27 10:37:48 crc kubenswrapper[4806]: I1127 10:37:48.638120 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rhbrc" event={"ID":"a3e71c8a-a010-49f6-9e79-c868ec3243d9","Type":"ContainerDied","Data":"b4a0eb7d0744a08010adc745dcfacc358029c1457f0b148d283dba18fe7bcd90"} Nov 27 10:37:48 crc kubenswrapper[4806]: I1127 10:37:48.641013 4806 generic.go:334] "Generic (PLEG): container finished" podID="679580af-7a44-485f-8b79-f79b702d85eb" containerID="85afbb8a4561fd7353250eb568d0df936c03dc313d30b9e1f4ce22eaaa1ae12e" exitCode=0 Nov 27 10:37:48 crc kubenswrapper[4806]: I1127 10:37:48.641090 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3a87-account-create-update-8zm2q" event={"ID":"679580af-7a44-485f-8b79-f79b702d85eb","Type":"ContainerDied","Data":"85afbb8a4561fd7353250eb568d0df936c03dc313d30b9e1f4ce22eaaa1ae12e"} Nov 27 10:37:48 crc kubenswrapper[4806]: I1127 10:37:48.646018 4806 generic.go:334] "Generic (PLEG): container finished" podID="f7143fe2-d8ae-4d55-a11c-edb06cf922f7" containerID="f6b40b2335c910e068c63dab1fac7deb316f1213e1f22109ff58b5a9dabe6b03" exitCode=0 Nov 27 10:37:48 crc kubenswrapper[4806]: I1127 10:37:48.646242 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xptpp" event={"ID":"f7143fe2-d8ae-4d55-a11c-edb06cf922f7","Type":"ContainerDied","Data":"f6b40b2335c910e068c63dab1fac7deb316f1213e1f22109ff58b5a9dabe6b03"} Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.526051 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xptpp" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.533499 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rhbrc" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.540154 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-f488s" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.550558 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3071-account-create-update-2g8jt" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.559650 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3a87-account-create-update-8zm2q" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.569684 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1c99-account-create-update-6mm4j" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.684581 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3071-account-create-update-2g8jt" event={"ID":"df636244-5318-44a8-857b-2c68af37b10a","Type":"ContainerDied","Data":"9dbefdb03453c954efa5edbfd746ee3de894e1c7c5e2de635eac0cb381a5b8f3"} Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.684643 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9dbefdb03453c954efa5edbfd746ee3de894e1c7c5e2de635eac0cb381a5b8f3" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.684643 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3071-account-create-update-2g8jt" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.685900 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-f488s" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.685884 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-f488s" event={"ID":"a9270118-f53f-4b38-98e6-03469db9302d","Type":"ContainerDied","Data":"d1a458be97a6aa38586471d651b06e63063a5c6a95ff77ea2b5e1e071099d9fd"} Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.686045 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1a458be97a6aa38586471d651b06e63063a5c6a95ff77ea2b5e1e071099d9fd" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.691448 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1c99-account-create-update-6mm4j" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.691685 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1c99-account-create-update-6mm4j" event={"ID":"82baaafd-2458-40d0-b859-8171f9a6db0d","Type":"ContainerDied","Data":"af0b08e34b17f0a2422e31bf51a54bb63701250b03bfa4fce101bfa1b663488d"} Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.691849 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af0b08e34b17f0a2422e31bf51a54bb63701250b03bfa4fce101bfa1b663488d" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.694462 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3a87-account-create-update-8zm2q" event={"ID":"679580af-7a44-485f-8b79-f79b702d85eb","Type":"ContainerDied","Data":"1a407f5950a84047e92f9112ab582c546a8061e82ed0b79216d7bf8231edccb4"} Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.694487 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a407f5950a84047e92f9112ab582c546a8061e82ed0b79216d7bf8231edccb4" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.694600 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3a87-account-create-update-8zm2q" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.696755 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2nlk\" (UniqueName: \"kubernetes.io/projected/df636244-5318-44a8-857b-2c68af37b10a-kube-api-access-h2nlk\") pod \"df636244-5318-44a8-857b-2c68af37b10a\" (UID: \"df636244-5318-44a8-857b-2c68af37b10a\") " Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.696794 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/679580af-7a44-485f-8b79-f79b702d85eb-operator-scripts\") pod \"679580af-7a44-485f-8b79-f79b702d85eb\" (UID: \"679580af-7a44-485f-8b79-f79b702d85eb\") " Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.696877 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5mxz\" (UniqueName: \"kubernetes.io/projected/82baaafd-2458-40d0-b859-8171f9a6db0d-kube-api-access-p5mxz\") pod \"82baaafd-2458-40d0-b859-8171f9a6db0d\" (UID: \"82baaafd-2458-40d0-b859-8171f9a6db0d\") " Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.696919 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsjqs\" (UniqueName: \"kubernetes.io/projected/a9270118-f53f-4b38-98e6-03469db9302d-kube-api-access-fsjqs\") pod \"a9270118-f53f-4b38-98e6-03469db9302d\" (UID: \"a9270118-f53f-4b38-98e6-03469db9302d\") " Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.696980 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7143fe2-d8ae-4d55-a11c-edb06cf922f7-operator-scripts\") pod \"f7143fe2-d8ae-4d55-a11c-edb06cf922f7\" (UID: \"f7143fe2-d8ae-4d55-a11c-edb06cf922f7\") " Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.697076 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3e71c8a-a010-49f6-9e79-c868ec3243d9-operator-scripts\") pod \"a3e71c8a-a010-49f6-9e79-c868ec3243d9\" (UID: \"a3e71c8a-a010-49f6-9e79-c868ec3243d9\") " Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.697171 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82baaafd-2458-40d0-b859-8171f9a6db0d-operator-scripts\") pod \"82baaafd-2458-40d0-b859-8171f9a6db0d\" (UID: \"82baaafd-2458-40d0-b859-8171f9a6db0d\") " Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.697319 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-242vj\" (UniqueName: \"kubernetes.io/projected/a3e71c8a-a010-49f6-9e79-c868ec3243d9-kube-api-access-242vj\") pod \"a3e71c8a-a010-49f6-9e79-c868ec3243d9\" (UID: \"a3e71c8a-a010-49f6-9e79-c868ec3243d9\") " Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.697377 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nb5dr\" (UniqueName: \"kubernetes.io/projected/679580af-7a44-485f-8b79-f79b702d85eb-kube-api-access-nb5dr\") pod \"679580af-7a44-485f-8b79-f79b702d85eb\" (UID: \"679580af-7a44-485f-8b79-f79b702d85eb\") " Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.697461 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df636244-5318-44a8-857b-2c68af37b10a-operator-scripts\") pod \"df636244-5318-44a8-857b-2c68af37b10a\" (UID: \"df636244-5318-44a8-857b-2c68af37b10a\") " Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.697578 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2drn\" (UniqueName: \"kubernetes.io/projected/f7143fe2-d8ae-4d55-a11c-edb06cf922f7-kube-api-access-m2drn\") pod \"f7143fe2-d8ae-4d55-a11c-edb06cf922f7\" (UID: \"f7143fe2-d8ae-4d55-a11c-edb06cf922f7\") " Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.697625 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9270118-f53f-4b38-98e6-03469db9302d-operator-scripts\") pod \"a9270118-f53f-4b38-98e6-03469db9302d\" (UID: \"a9270118-f53f-4b38-98e6-03469db9302d\") " Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.699123 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/679580af-7a44-485f-8b79-f79b702d85eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "679580af-7a44-485f-8b79-f79b702d85eb" (UID: "679580af-7a44-485f-8b79-f79b702d85eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.700328 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82baaafd-2458-40d0-b859-8171f9a6db0d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "82baaafd-2458-40d0-b859-8171f9a6db0d" (UID: "82baaafd-2458-40d0-b859-8171f9a6db0d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.700629 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9270118-f53f-4b38-98e6-03469db9302d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a9270118-f53f-4b38-98e6-03469db9302d" (UID: "a9270118-f53f-4b38-98e6-03469db9302d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.701397 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3e71c8a-a010-49f6-9e79-c868ec3243d9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a3e71c8a-a010-49f6-9e79-c868ec3243d9" (UID: "a3e71c8a-a010-49f6-9e79-c868ec3243d9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.701540 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7143fe2-d8ae-4d55-a11c-edb06cf922f7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f7143fe2-d8ae-4d55-a11c-edb06cf922f7" (UID: "f7143fe2-d8ae-4d55-a11c-edb06cf922f7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.702612 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df636244-5318-44a8-857b-2c68af37b10a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "df636244-5318-44a8-857b-2c68af37b10a" (UID: "df636244-5318-44a8-857b-2c68af37b10a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.702633 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xptpp" event={"ID":"f7143fe2-d8ae-4d55-a11c-edb06cf922f7","Type":"ContainerDied","Data":"af94e764c9a902b653eb47642fffd8ce8208bf699bb6b2690f428d5e6be2aa79"} Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.702703 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xptpp" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.702702 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af94e764c9a902b653eb47642fffd8ce8208bf699bb6b2690f428d5e6be2aa79" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.705330 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3e71c8a-a010-49f6-9e79-c868ec3243d9-kube-api-access-242vj" (OuterVolumeSpecName: "kube-api-access-242vj") pod "a3e71c8a-a010-49f6-9e79-c868ec3243d9" (UID: "a3e71c8a-a010-49f6-9e79-c868ec3243d9"). InnerVolumeSpecName "kube-api-access-242vj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.706033 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9270118-f53f-4b38-98e6-03469db9302d-kube-api-access-fsjqs" (OuterVolumeSpecName: "kube-api-access-fsjqs") pod "a9270118-f53f-4b38-98e6-03469db9302d" (UID: "a9270118-f53f-4b38-98e6-03469db9302d"). InnerVolumeSpecName "kube-api-access-fsjqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.706851 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7143fe2-d8ae-4d55-a11c-edb06cf922f7-kube-api-access-m2drn" (OuterVolumeSpecName: "kube-api-access-m2drn") pod "f7143fe2-d8ae-4d55-a11c-edb06cf922f7" (UID: "f7143fe2-d8ae-4d55-a11c-edb06cf922f7"). InnerVolumeSpecName "kube-api-access-m2drn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.707219 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/679580af-7a44-485f-8b79-f79b702d85eb-kube-api-access-nb5dr" (OuterVolumeSpecName: "kube-api-access-nb5dr") pod "679580af-7a44-485f-8b79-f79b702d85eb" (UID: "679580af-7a44-485f-8b79-f79b702d85eb"). InnerVolumeSpecName "kube-api-access-nb5dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.707707 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rhbrc" event={"ID":"a3e71c8a-a010-49f6-9e79-c868ec3243d9","Type":"ContainerDied","Data":"d23551eb8426ee5dbefe6bc762d172b8c2ecc3d5fd8aeceed84cc17ae6ef6f88"} Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.707846 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rhbrc" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.708508 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d23551eb8426ee5dbefe6bc762d172b8c2ecc3d5fd8aeceed84cc17ae6ef6f88" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.708743 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df636244-5318-44a8-857b-2c68af37b10a-kube-api-access-h2nlk" (OuterVolumeSpecName: "kube-api-access-h2nlk") pod "df636244-5318-44a8-857b-2c68af37b10a" (UID: "df636244-5318-44a8-857b-2c68af37b10a"). InnerVolumeSpecName "kube-api-access-h2nlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.711536 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82baaafd-2458-40d0-b859-8171f9a6db0d-kube-api-access-p5mxz" (OuterVolumeSpecName: "kube-api-access-p5mxz") pod "82baaafd-2458-40d0-b859-8171f9a6db0d" (UID: "82baaafd-2458-40d0-b859-8171f9a6db0d"). InnerVolumeSpecName "kube-api-access-p5mxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.799910 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df636244-5318-44a8-857b-2c68af37b10a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.799953 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2drn\" (UniqueName: \"kubernetes.io/projected/f7143fe2-d8ae-4d55-a11c-edb06cf922f7-kube-api-access-m2drn\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.799963 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9270118-f53f-4b38-98e6-03469db9302d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.799977 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2nlk\" (UniqueName: \"kubernetes.io/projected/df636244-5318-44a8-857b-2c68af37b10a-kube-api-access-h2nlk\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.799987 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/679580af-7a44-485f-8b79-f79b702d85eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.799999 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5mxz\" (UniqueName: \"kubernetes.io/projected/82baaafd-2458-40d0-b859-8171f9a6db0d-kube-api-access-p5mxz\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.800007 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsjqs\" (UniqueName: \"kubernetes.io/projected/a9270118-f53f-4b38-98e6-03469db9302d-kube-api-access-fsjqs\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.800017 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7143fe2-d8ae-4d55-a11c-edb06cf922f7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.800027 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3e71c8a-a010-49f6-9e79-c868ec3243d9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.800036 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82baaafd-2458-40d0-b859-8171f9a6db0d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.800047 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-242vj\" (UniqueName: \"kubernetes.io/projected/a3e71c8a-a010-49f6-9e79-c868ec3243d9-kube-api-access-242vj\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:52 crc kubenswrapper[4806]: I1127 10:37:52.800057 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nb5dr\" (UniqueName: \"kubernetes.io/projected/679580af-7a44-485f-8b79-f79b702d85eb-kube-api-access-nb5dr\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:53 crc kubenswrapper[4806]: I1127 10:37:53.717090 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6vcz5" event={"ID":"c7484e21-2c80-49d6-a9ec-6489bc40ec4f","Type":"ContainerStarted","Data":"35b4ed1f1f67e783c6ecb121e37e9be0153a83bf8984f252587b15c9d553781c"} Nov 27 10:37:53 crc kubenswrapper[4806]: I1127 10:37:53.733468 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-6vcz5" podStartSLOduration=2.5206759229999998 podStartE2EDuration="10.733434733s" podCreationTimestamp="2025-11-27 10:37:43 +0000 UTC" firstStartedPulling="2025-11-27 10:37:44.955272974 +0000 UTC m=+969.541863728" lastFinishedPulling="2025-11-27 10:37:53.168031764 +0000 UTC m=+977.754622538" observedRunningTime="2025-11-27 10:37:53.732788634 +0000 UTC m=+978.319379398" watchObservedRunningTime="2025-11-27 10:37:53.733434733 +0000 UTC m=+978.320025497" Nov 27 10:37:56 crc kubenswrapper[4806]: I1127 10:37:56.746986 4806 generic.go:334] "Generic (PLEG): container finished" podID="c7484e21-2c80-49d6-a9ec-6489bc40ec4f" containerID="35b4ed1f1f67e783c6ecb121e37e9be0153a83bf8984f252587b15c9d553781c" exitCode=0 Nov 27 10:37:56 crc kubenswrapper[4806]: I1127 10:37:56.747069 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6vcz5" event={"ID":"c7484e21-2c80-49d6-a9ec-6489bc40ec4f","Type":"ContainerDied","Data":"35b4ed1f1f67e783c6ecb121e37e9be0153a83bf8984f252587b15c9d553781c"} Nov 27 10:37:57 crc kubenswrapper[4806]: I1127 10:37:57.762956 4806 generic.go:334] "Generic (PLEG): container finished" podID="9aefc973-2bd6-449e-a7e7-afd6a7915c7a" containerID="856d1cbfd076b861d6ec0311109278411b8f03ca997a7f6e4b585020b52cb83f" exitCode=0 Nov 27 10:37:57 crc kubenswrapper[4806]: I1127 10:37:57.763072 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h8mc6" event={"ID":"9aefc973-2bd6-449e-a7e7-afd6a7915c7a","Type":"ContainerDied","Data":"856d1cbfd076b861d6ec0311109278411b8f03ca997a7f6e4b585020b52cb83f"} Nov 27 10:37:58 crc kubenswrapper[4806]: I1127 10:37:58.145692 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6vcz5" Nov 27 10:37:58 crc kubenswrapper[4806]: I1127 10:37:58.296736 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-config-data\") pod \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\" (UID: \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\") " Nov 27 10:37:58 crc kubenswrapper[4806]: I1127 10:37:58.296827 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg2xq\" (UniqueName: \"kubernetes.io/projected/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-kube-api-access-qg2xq\") pod \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\" (UID: \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\") " Nov 27 10:37:58 crc kubenswrapper[4806]: I1127 10:37:58.296882 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-combined-ca-bundle\") pod \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\" (UID: \"c7484e21-2c80-49d6-a9ec-6489bc40ec4f\") " Nov 27 10:37:58 crc kubenswrapper[4806]: I1127 10:37:58.305624 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-kube-api-access-qg2xq" (OuterVolumeSpecName: "kube-api-access-qg2xq") pod "c7484e21-2c80-49d6-a9ec-6489bc40ec4f" (UID: "c7484e21-2c80-49d6-a9ec-6489bc40ec4f"). InnerVolumeSpecName "kube-api-access-qg2xq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:58 crc kubenswrapper[4806]: I1127 10:37:58.333549 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7484e21-2c80-49d6-a9ec-6489bc40ec4f" (UID: "c7484e21-2c80-49d6-a9ec-6489bc40ec4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:37:58 crc kubenswrapper[4806]: I1127 10:37:58.343455 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-config-data" (OuterVolumeSpecName: "config-data") pod "c7484e21-2c80-49d6-a9ec-6489bc40ec4f" (UID: "c7484e21-2c80-49d6-a9ec-6489bc40ec4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:37:58 crc kubenswrapper[4806]: I1127 10:37:58.399156 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg2xq\" (UniqueName: \"kubernetes.io/projected/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-kube-api-access-qg2xq\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:58 crc kubenswrapper[4806]: I1127 10:37:58.399189 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:58 crc kubenswrapper[4806]: I1127 10:37:58.399201 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7484e21-2c80-49d6-a9ec-6489bc40ec4f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:58 crc kubenswrapper[4806]: I1127 10:37:58.773817 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6vcz5" event={"ID":"c7484e21-2c80-49d6-a9ec-6489bc40ec4f","Type":"ContainerDied","Data":"d286b5f121341c8d7228c373cd17401952e4de446db8a52b75b35a2d5735fb8d"} Nov 27 10:37:58 crc kubenswrapper[4806]: I1127 10:37:58.774170 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d286b5f121341c8d7228c373cd17401952e4de446db8a52b75b35a2d5735fb8d" Nov 27 10:37:58 crc kubenswrapper[4806]: I1127 10:37:58.773989 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6vcz5" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091115 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-665mv"] Nov 27 10:37:59 crc kubenswrapper[4806]: E1127 10:37:59.091504 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df636244-5318-44a8-857b-2c68af37b10a" containerName="mariadb-account-create-update" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091518 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="df636244-5318-44a8-857b-2c68af37b10a" containerName="mariadb-account-create-update" Nov 27 10:37:59 crc kubenswrapper[4806]: E1127 10:37:59.091527 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="679580af-7a44-485f-8b79-f79b702d85eb" containerName="mariadb-account-create-update" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091533 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="679580af-7a44-485f-8b79-f79b702d85eb" containerName="mariadb-account-create-update" Nov 27 10:37:59 crc kubenswrapper[4806]: E1127 10:37:59.091542 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9270118-f53f-4b38-98e6-03469db9302d" containerName="mariadb-database-create" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091548 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9270118-f53f-4b38-98e6-03469db9302d" containerName="mariadb-database-create" Nov 27 10:37:59 crc kubenswrapper[4806]: E1127 10:37:59.091562 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82baaafd-2458-40d0-b859-8171f9a6db0d" containerName="mariadb-account-create-update" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091568 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="82baaafd-2458-40d0-b859-8171f9a6db0d" containerName="mariadb-account-create-update" Nov 27 10:37:59 crc kubenswrapper[4806]: E1127 10:37:59.091577 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee90ab60-6e5b-44f5-9445-f874090f7c26" containerName="ovn-config" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091583 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee90ab60-6e5b-44f5-9445-f874090f7c26" containerName="ovn-config" Nov 27 10:37:59 crc kubenswrapper[4806]: E1127 10:37:59.091595 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3e71c8a-a010-49f6-9e79-c868ec3243d9" containerName="mariadb-database-create" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091601 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3e71c8a-a010-49f6-9e79-c868ec3243d9" containerName="mariadb-database-create" Nov 27 10:37:59 crc kubenswrapper[4806]: E1127 10:37:59.091611 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7484e21-2c80-49d6-a9ec-6489bc40ec4f" containerName="keystone-db-sync" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091617 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7484e21-2c80-49d6-a9ec-6489bc40ec4f" containerName="keystone-db-sync" Nov 27 10:37:59 crc kubenswrapper[4806]: E1127 10:37:59.091640 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7143fe2-d8ae-4d55-a11c-edb06cf922f7" containerName="mariadb-database-create" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091646 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7143fe2-d8ae-4d55-a11c-edb06cf922f7" containerName="mariadb-database-create" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091799 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3e71c8a-a010-49f6-9e79-c868ec3243d9" containerName="mariadb-database-create" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091810 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee90ab60-6e5b-44f5-9445-f874090f7c26" containerName="ovn-config" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091818 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9270118-f53f-4b38-98e6-03469db9302d" containerName="mariadb-database-create" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091826 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="df636244-5318-44a8-857b-2c68af37b10a" containerName="mariadb-account-create-update" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091836 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7484e21-2c80-49d6-a9ec-6489bc40ec4f" containerName="keystone-db-sync" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091846 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="82baaafd-2458-40d0-b859-8171f9a6db0d" containerName="mariadb-account-create-update" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091855 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7143fe2-d8ae-4d55-a11c-edb06cf922f7" containerName="mariadb-database-create" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.091862 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="679580af-7a44-485f-8b79-f79b702d85eb" containerName="mariadb-account-create-update" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.092453 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.097154 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.097883 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.098004 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.098106 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxgbr" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.098265 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.125678 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-pnvmt"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.127044 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.155495 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-665mv"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.193164 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-pnvmt"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.226249 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-credential-keys\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.226318 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sfk7\" (UniqueName: \"kubernetes.io/projected/7dad6a31-52cc-4921-b585-2a69c9ebc607-kube-api-access-4sfk7\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.226421 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-ovsdbserver-nb\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.226838 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-ovsdbserver-sb\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.226895 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-scripts\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.226976 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-combined-ca-bundle\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.227039 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-fernet-keys\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.227105 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdjm7\" (UniqueName: \"kubernetes.io/projected/12d9875d-e81f-425b-9821-73395c5c1632-kube-api-access-vdjm7\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.227153 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-config-data\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.227256 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-config\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.227297 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-dns-svc\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.328932 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-config\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.328987 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-dns-svc\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.329037 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-credential-keys\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.329072 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sfk7\" (UniqueName: \"kubernetes.io/projected/7dad6a31-52cc-4921-b585-2a69c9ebc607-kube-api-access-4sfk7\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.329120 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-ovsdbserver-nb\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.329152 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-ovsdbserver-sb\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.329170 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-scripts\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.329197 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-combined-ca-bundle\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.329223 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-fernet-keys\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.329268 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdjm7\" (UniqueName: \"kubernetes.io/projected/12d9875d-e81f-425b-9821-73395c5c1632-kube-api-access-vdjm7\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.329331 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-config-data\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.330805 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-ovsdbserver-nb\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.331608 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-ovsdbserver-sb\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.332060 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-config\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.333123 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-dns-svc\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.345013 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-combined-ca-bundle\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.345697 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-fernet-keys\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.346041 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-config-data\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.346433 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-scripts\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.353172 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-credential-keys\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.367883 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sfk7\" (UniqueName: \"kubernetes.io/projected/7dad6a31-52cc-4921-b585-2a69c9ebc607-kube-api-access-4sfk7\") pod \"keystone-bootstrap-665mv\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.390475 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdjm7\" (UniqueName: \"kubernetes.io/projected/12d9875d-e81f-425b-9821-73395c5c1632-kube-api-access-vdjm7\") pod \"dnsmasq-dns-66fbd85b65-pnvmt\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.421521 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-cwqkj"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.448128 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.458670 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-665mv" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.460196 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5jcsw" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.460652 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.460816 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.470623 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.504495 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-cwqkj"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.506048 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.550982 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnctc\" (UniqueName: \"kubernetes.io/projected/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-kube-api-access-nnctc\") pod \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.551326 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-config-data\") pod \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.566797 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-combined-ca-bundle\") pod \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.574954 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-db-sync-config-data\") pod \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\" (UID: \"9aefc973-2bd6-449e-a7e7-afd6a7915c7a\") " Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.575837 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-scripts\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.576074 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-db-sync-config-data\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.584434 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/552d400b-d5b2-40c8-9654-bfc70ab5bab6-etc-machine-id\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.584875 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnt7d\" (UniqueName: \"kubernetes.io/projected/552d400b-d5b2-40c8-9654-bfc70ab5bab6-kube-api-access-lnt7d\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.585092 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-config-data\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.585272 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-combined-ca-bundle\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.599343 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9aefc973-2bd6-449e-a7e7-afd6a7915c7a" (UID: "9aefc973-2bd6-449e-a7e7-afd6a7915c7a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.607008 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-kube-api-access-nnctc" (OuterVolumeSpecName: "kube-api-access-nnctc") pod "9aefc973-2bd6-449e-a7e7-afd6a7915c7a" (UID: "9aefc973-2bd6-449e-a7e7-afd6a7915c7a"). InnerVolumeSpecName "kube-api-access-nnctc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.681448 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-9vkqx"] Nov 27 10:37:59 crc kubenswrapper[4806]: E1127 10:37:59.681976 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aefc973-2bd6-449e-a7e7-afd6a7915c7a" containerName="glance-db-sync" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.681997 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aefc973-2bd6-449e-a7e7-afd6a7915c7a" containerName="glance-db-sync" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.682194 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aefc973-2bd6-449e-a7e7-afd6a7915c7a" containerName="glance-db-sync" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.682967 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9vkqx" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.686823 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/552d400b-d5b2-40c8-9654-bfc70ab5bab6-etc-machine-id\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.686885 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnt7d\" (UniqueName: \"kubernetes.io/projected/552d400b-d5b2-40c8-9654-bfc70ab5bab6-kube-api-access-lnt7d\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.686923 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-config-data\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.686955 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-combined-ca-bundle\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.686989 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-scripts\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.687023 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-db-sync-config-data\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.687070 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnctc\" (UniqueName: \"kubernetes.io/projected/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-kube-api-access-nnctc\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.687081 4806 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.690052 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.692879 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/552d400b-d5b2-40c8-9654-bfc70ab5bab6-etc-machine-id\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.695758 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hdrsj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.737597 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-config-data\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.737817 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-db-sync-config-data\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.738174 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-combined-ca-bundle\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.738750 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnt7d\" (UniqueName: \"kubernetes.io/projected/552d400b-d5b2-40c8-9654-bfc70ab5bab6-kube-api-access-lnt7d\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.739447 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-scripts\") pod \"cinder-db-sync-cwqkj\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.750855 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9vkqx"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.769502 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-h7fbq"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.773095 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-h7fbq" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.777493 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.779499 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9aefc973-2bd6-449e-a7e7-afd6a7915c7a" (UID: "9aefc973-2bd6-449e-a7e7-afd6a7915c7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.792557 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b83e17-71ca-4f23-ad17-950a5de85a12-combined-ca-bundle\") pod \"barbican-db-sync-9vkqx\" (UID: \"11b83e17-71ca-4f23-ad17-950a5de85a12\") " pod="openstack/barbican-db-sync-9vkqx" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.792647 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11b83e17-71ca-4f23-ad17-950a5de85a12-db-sync-config-data\") pod \"barbican-db-sync-9vkqx\" (UID: \"11b83e17-71ca-4f23-ad17-950a5de85a12\") " pod="openstack/barbican-db-sync-9vkqx" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.792708 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzbjf\" (UniqueName: \"kubernetes.io/projected/11b83e17-71ca-4f23-ad17-950a5de85a12-kube-api-access-pzbjf\") pod \"barbican-db-sync-9vkqx\" (UID: \"11b83e17-71ca-4f23-ad17-950a5de85a12\") " pod="openstack/barbican-db-sync-9vkqx" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.792747 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.795914 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-h7fbq"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.797378 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-config-data" (OuterVolumeSpecName: "config-data") pod "9aefc973-2bd6-449e-a7e7-afd6a7915c7a" (UID: "9aefc973-2bd6-449e-a7e7-afd6a7915c7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.797860 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.798125 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.798448 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9v77k" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.810304 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-pnvmt"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.819365 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h8mc6" event={"ID":"9aefc973-2bd6-449e-a7e7-afd6a7915c7a","Type":"ContainerDied","Data":"ecb4fbd1fa88d7c7b6953b28a65d3cd1d5b98f5d3c1e77212ff7b4b4c3b39e95"} Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.819509 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecb4fbd1fa88d7c7b6953b28a65d3cd1d5b98f5d3c1e77212ff7b4b4c3b39e95" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.819642 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h8mc6" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.887497 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-8mddk"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.892447 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.893808 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11b83e17-71ca-4f23-ad17-950a5de85a12-db-sync-config-data\") pod \"barbican-db-sync-9vkqx\" (UID: \"11b83e17-71ca-4f23-ad17-950a5de85a12\") " pod="openstack/barbican-db-sync-9vkqx" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.893875 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzbjf\" (UniqueName: \"kubernetes.io/projected/11b83e17-71ca-4f23-ad17-950a5de85a12-kube-api-access-pzbjf\") pod \"barbican-db-sync-9vkqx\" (UID: \"11b83e17-71ca-4f23-ad17-950a5de85a12\") " pod="openstack/barbican-db-sync-9vkqx" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.893927 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv2qr\" (UniqueName: \"kubernetes.io/projected/94086dd9-9626-49a4-ad72-af183dc0a161-kube-api-access-kv2qr\") pod \"neutron-db-sync-h7fbq\" (UID: \"94086dd9-9626-49a4-ad72-af183dc0a161\") " pod="openstack/neutron-db-sync-h7fbq" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.893953 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/94086dd9-9626-49a4-ad72-af183dc0a161-config\") pod \"neutron-db-sync-h7fbq\" (UID: \"94086dd9-9626-49a4-ad72-af183dc0a161\") " pod="openstack/neutron-db-sync-h7fbq" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.893990 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b83e17-71ca-4f23-ad17-950a5de85a12-combined-ca-bundle\") pod \"barbican-db-sync-9vkqx\" (UID: \"11b83e17-71ca-4f23-ad17-950a5de85a12\") " pod="openstack/barbican-db-sync-9vkqx" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.894011 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94086dd9-9626-49a4-ad72-af183dc0a161-combined-ca-bundle\") pod \"neutron-db-sync-h7fbq\" (UID: \"94086dd9-9626-49a4-ad72-af183dc0a161\") " pod="openstack/neutron-db-sync-h7fbq" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.894075 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aefc973-2bd6-449e-a7e7-afd6a7915c7a-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.902643 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b83e17-71ca-4f23-ad17-950a5de85a12-combined-ca-bundle\") pod \"barbican-db-sync-9vkqx\" (UID: \"11b83e17-71ca-4f23-ad17-950a5de85a12\") " pod="openstack/barbican-db-sync-9vkqx" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.912582 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-hz79k"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.912797 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11b83e17-71ca-4f23-ad17-950a5de85a12-db-sync-config-data\") pod \"barbican-db-sync-9vkqx\" (UID: \"11b83e17-71ca-4f23-ad17-950a5de85a12\") " pod="openstack/barbican-db-sync-9vkqx" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.913725 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hz79k" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.918315 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-8mddk"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.924016 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzbjf\" (UniqueName: \"kubernetes.io/projected/11b83e17-71ca-4f23-ad17-950a5de85a12-kube-api-access-pzbjf\") pod \"barbican-db-sync-9vkqx\" (UID: \"11b83e17-71ca-4f23-ad17-950a5de85a12\") " pod="openstack/barbican-db-sync-9vkqx" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.932442 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hz79k"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.933693 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.933866 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.934124 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-sfrbs" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.976125 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.978174 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.981752 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.981906 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.990907 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.996674 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-combined-ca-bundle\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.996724 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94086dd9-9626-49a4-ad72-af183dc0a161-combined-ca-bundle\") pod \"neutron-db-sync-h7fbq\" (UID: \"94086dd9-9626-49a4-ad72-af183dc0a161\") " pod="openstack/neutron-db-sync-h7fbq" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.996751 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-config\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.996776 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-scripts\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.996802 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b289081e-bb54-4046-9e62-37e47cacca99-logs\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.996838 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlrgx\" (UniqueName: \"kubernetes.io/projected/734efb9b-df1d-4f4f-ba47-7df13a997e3e-kube-api-access-zlrgx\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.996874 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpgc2\" (UniqueName: \"kubernetes.io/projected/b289081e-bb54-4046-9e62-37e47cacca99-kube-api-access-zpgc2\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.996904 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.996923 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-dns-svc\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.996947 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.996966 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv2qr\" (UniqueName: \"kubernetes.io/projected/94086dd9-9626-49a4-ad72-af183dc0a161-kube-api-access-kv2qr\") pod \"neutron-db-sync-h7fbq\" (UID: \"94086dd9-9626-49a4-ad72-af183dc0a161\") " pod="openstack/neutron-db-sync-h7fbq" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.996987 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-config-data\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:37:59 crc kubenswrapper[4806]: I1127 10:37:59.997005 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/94086dd9-9626-49a4-ad72-af183dc0a161-config\") pod \"neutron-db-sync-h7fbq\" (UID: \"94086dd9-9626-49a4-ad72-af183dc0a161\") " pod="openstack/neutron-db-sync-h7fbq" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.003349 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94086dd9-9626-49a4-ad72-af183dc0a161-combined-ca-bundle\") pod \"neutron-db-sync-h7fbq\" (UID: \"94086dd9-9626-49a4-ad72-af183dc0a161\") " pod="openstack/neutron-db-sync-h7fbq" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.014674 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/94086dd9-9626-49a4-ad72-af183dc0a161-config\") pod \"neutron-db-sync-h7fbq\" (UID: \"94086dd9-9626-49a4-ad72-af183dc0a161\") " pod="openstack/neutron-db-sync-h7fbq" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.035326 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv2qr\" (UniqueName: \"kubernetes.io/projected/94086dd9-9626-49a4-ad72-af183dc0a161-kube-api-access-kv2qr\") pod \"neutron-db-sync-h7fbq\" (UID: \"94086dd9-9626-49a4-ad72-af183dc0a161\") " pod="openstack/neutron-db-sync-h7fbq" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.046414 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9vkqx" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099275 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05ff370-09c4-45c1-b101-a6b277a34d3d-run-httpd\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099359 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099383 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-dns-svc\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099418 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099445 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099470 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-config-data\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099494 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-scripts\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099520 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-config-data\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099550 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099571 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05ff370-09c4-45c1-b101-a6b277a34d3d-log-httpd\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099595 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-combined-ca-bundle\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099620 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znq84\" (UniqueName: \"kubernetes.io/projected/d05ff370-09c4-45c1-b101-a6b277a34d3d-kube-api-access-znq84\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099638 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-config\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099666 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-scripts\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099695 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b289081e-bb54-4046-9e62-37e47cacca99-logs\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099736 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlrgx\" (UniqueName: \"kubernetes.io/projected/734efb9b-df1d-4f4f-ba47-7df13a997e3e-kube-api-access-zlrgx\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.099780 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpgc2\" (UniqueName: \"kubernetes.io/projected/b289081e-bb54-4046-9e62-37e47cacca99-kube-api-access-zpgc2\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.101276 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.101754 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b289081e-bb54-4046-9e62-37e47cacca99-logs\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.102475 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.102798 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-config\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.103162 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-dns-svc\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.127964 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-scripts\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.129292 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-h7fbq" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.148788 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-config-data\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.153150 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-combined-ca-bundle\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.166559 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpgc2\" (UniqueName: \"kubernetes.io/projected/b289081e-bb54-4046-9e62-37e47cacca99-kube-api-access-zpgc2\") pod \"placement-db-sync-hz79k\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " pod="openstack/placement-db-sync-hz79k" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.168121 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlrgx\" (UniqueName: \"kubernetes.io/projected/734efb9b-df1d-4f4f-ba47-7df13a997e3e-kube-api-access-zlrgx\") pod \"dnsmasq-dns-6bf59f66bf-8mddk\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.204481 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05ff370-09c4-45c1-b101-a6b277a34d3d-run-httpd\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.204552 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.204594 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-scripts\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.204611 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-config-data\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.204629 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.204646 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05ff370-09c4-45c1-b101-a6b277a34d3d-log-httpd\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.204683 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znq84\" (UniqueName: \"kubernetes.io/projected/d05ff370-09c4-45c1-b101-a6b277a34d3d-kube-api-access-znq84\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.212783 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-pnvmt"] Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.218395 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05ff370-09c4-45c1-b101-a6b277a34d3d-run-httpd\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.225017 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05ff370-09c4-45c1-b101-a6b277a34d3d-log-httpd\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.233731 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-scripts\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.238318 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.238845 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-config-data\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.255133 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.256152 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.270465 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hz79k" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.282032 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znq84\" (UniqueName: \"kubernetes.io/projected/d05ff370-09c4-45c1-b101-a6b277a34d3d-kube-api-access-znq84\") pod \"ceilometer-0\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.321425 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.368667 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-8mddk"] Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.386251 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-665mv"] Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.393779 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw"] Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.395148 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.472468 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw"] Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.518331 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9rln\" (UniqueName: \"kubernetes.io/projected/68ebcee0-f362-4953-8775-1add49fb4add-kube-api-access-p9rln\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.518379 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-config\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.518437 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.518484 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.518509 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.664301 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.664408 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.664444 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.664519 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9rln\" (UniqueName: \"kubernetes.io/projected/68ebcee0-f362-4953-8775-1add49fb4add-kube-api-access-p9rln\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.664538 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-config\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.666656 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-config\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.678921 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.685193 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.697467 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.716286 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9rln\" (UniqueName: \"kubernetes.io/projected/68ebcee0-f362-4953-8775-1add49fb4add-kube-api-access-p9rln\") pod \"dnsmasq-dns-5b6dbdb6f5-8l2bw\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.717437 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.891148 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-665mv" event={"ID":"7dad6a31-52cc-4921-b585-2a69c9ebc607","Type":"ContainerStarted","Data":"e9b58bdd1816684d4ec0778b4ba938d63bf26ea7545af2ef90f7f7e43ab3aa91"} Nov 27 10:38:00 crc kubenswrapper[4806]: I1127 10:38:00.925208 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" event={"ID":"12d9875d-e81f-425b-9821-73395c5c1632","Type":"ContainerStarted","Data":"e4b4b0b05d709f8613968c546f65f049a51b2b6bcac26895a40aa526c0ba9fea"} Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.084358 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-cwqkj"] Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.104973 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9vkqx"] Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.433759 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-h7fbq"] Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.504106 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-8mddk"] Nov 27 10:38:01 crc kubenswrapper[4806]: E1127 10:38:01.618839 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12d9875d_e81f_425b_9821_73395c5c1632.slice/crio-b141e7d6e9c2fa98de9d080ba8feb89f941c9f7289c1e1e24a206ce2c62d6d20.scope\": RecentStats: unable to find data in memory cache]" Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.649239 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.790628 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw"] Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.802148 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hz79k"] Nov 27 10:38:01 crc kubenswrapper[4806]: W1127 10:38:01.804135 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68ebcee0_f362_4953_8775_1add49fb4add.slice/crio-82d11378477f383061a1a868603a3737c42b192c785ff61fc8cba2df1a656ef7 WatchSource:0}: Error finding container 82d11378477f383061a1a868603a3737c42b192c785ff61fc8cba2df1a656ef7: Status 404 returned error can't find the container with id 82d11378477f383061a1a868603a3737c42b192c785ff61fc8cba2df1a656ef7 Nov 27 10:38:01 crc kubenswrapper[4806]: W1127 10:38:01.811432 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb289081e_bb54_4046_9e62_37e47cacca99.slice/crio-e5ea685ebd5fd28b91208d64aef30c5681931774720f7cb0043f46d8afcb6998 WatchSource:0}: Error finding container e5ea685ebd5fd28b91208d64aef30c5681931774720f7cb0043f46d8afcb6998: Status 404 returned error can't find the container with id e5ea685ebd5fd28b91208d64aef30c5681931774720f7cb0043f46d8afcb6998 Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.955942 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" event={"ID":"734efb9b-df1d-4f4f-ba47-7df13a997e3e","Type":"ContainerStarted","Data":"1049580d8d8a5a36485318937ec6c97cfc671b49451795ec768c9919c961fa65"} Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.970270 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9vkqx" event={"ID":"11b83e17-71ca-4f23-ad17-950a5de85a12","Type":"ContainerStarted","Data":"56f7d94746a17b9cd17a006fddd49da69b3e6c38fde001caea0baf9ab370c0f5"} Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.973000 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05ff370-09c4-45c1-b101-a6b277a34d3d","Type":"ContainerStarted","Data":"091208a112c811eb72cbf388928ae445197a04ac2d0fd48f1ca61a17ffe2b085"} Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.991129 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" event={"ID":"68ebcee0-f362-4953-8775-1add49fb4add","Type":"ContainerStarted","Data":"82d11378477f383061a1a868603a3737c42b192c785ff61fc8cba2df1a656ef7"} Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.992691 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hz79k" event={"ID":"b289081e-bb54-4046-9e62-37e47cacca99","Type":"ContainerStarted","Data":"e5ea685ebd5fd28b91208d64aef30c5681931774720f7cb0043f46d8afcb6998"} Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.994194 4806 generic.go:334] "Generic (PLEG): container finished" podID="12d9875d-e81f-425b-9821-73395c5c1632" containerID="b141e7d6e9c2fa98de9d080ba8feb89f941c9f7289c1e1e24a206ce2c62d6d20" exitCode=0 Nov 27 10:38:01 crc kubenswrapper[4806]: I1127 10:38:01.994313 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" event={"ID":"12d9875d-e81f-425b-9821-73395c5c1632","Type":"ContainerDied","Data":"b141e7d6e9c2fa98de9d080ba8feb89f941c9f7289c1e1e24a206ce2c62d6d20"} Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.002632 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-665mv" event={"ID":"7dad6a31-52cc-4921-b585-2a69c9ebc607","Type":"ContainerStarted","Data":"b5f34a455734a87a2d12b56355e08ede8ebdacaade0e860817dc1ea7ce12d66b"} Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.007425 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cwqkj" event={"ID":"552d400b-d5b2-40c8-9654-bfc70ab5bab6","Type":"ContainerStarted","Data":"7b53b5c38a61c2b7430cfbff74b42faaeec94ad45cf3647384d4c18dfd638a69"} Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.031862 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-h7fbq" event={"ID":"94086dd9-9626-49a4-ad72-af183dc0a161","Type":"ContainerStarted","Data":"d2a5eda0802027071d1e68bf8b7f9de6516a5df948f774197491684a44f05028"} Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.031913 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-h7fbq" event={"ID":"94086dd9-9626-49a4-ad72-af183dc0a161","Type":"ContainerStarted","Data":"6a43ed6818a88daf6385e875eae5e84d01e9f8a360ee97893b3358ffb712478c"} Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.053890 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-665mv" podStartSLOduration=3.053868376 podStartE2EDuration="3.053868376s" podCreationTimestamp="2025-11-27 10:37:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:38:02.047469718 +0000 UTC m=+986.634060482" watchObservedRunningTime="2025-11-27 10:38:02.053868376 +0000 UTC m=+986.640459140" Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.079780 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-h7fbq" podStartSLOduration=3.0797616740000002 podStartE2EDuration="3.079761674s" podCreationTimestamp="2025-11-27 10:37:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:38:02.078418546 +0000 UTC m=+986.665009310" watchObservedRunningTime="2025-11-27 10:38:02.079761674 +0000 UTC m=+986.666352438" Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.471738 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.647954 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdjm7\" (UniqueName: \"kubernetes.io/projected/12d9875d-e81f-425b-9821-73395c5c1632-kube-api-access-vdjm7\") pod \"12d9875d-e81f-425b-9821-73395c5c1632\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.648042 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-dns-svc\") pod \"12d9875d-e81f-425b-9821-73395c5c1632\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.648119 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-ovsdbserver-sb\") pod \"12d9875d-e81f-425b-9821-73395c5c1632\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.648309 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-ovsdbserver-nb\") pod \"12d9875d-e81f-425b-9821-73395c5c1632\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.648415 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-config\") pod \"12d9875d-e81f-425b-9821-73395c5c1632\" (UID: \"12d9875d-e81f-425b-9821-73395c5c1632\") " Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.678305 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12d9875d-e81f-425b-9821-73395c5c1632-kube-api-access-vdjm7" (OuterVolumeSpecName: "kube-api-access-vdjm7") pod "12d9875d-e81f-425b-9821-73395c5c1632" (UID: "12d9875d-e81f-425b-9821-73395c5c1632"). InnerVolumeSpecName "kube-api-access-vdjm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.681949 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-config" (OuterVolumeSpecName: "config") pod "12d9875d-e81f-425b-9821-73395c5c1632" (UID: "12d9875d-e81f-425b-9821-73395c5c1632"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.689654 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "12d9875d-e81f-425b-9821-73395c5c1632" (UID: "12d9875d-e81f-425b-9821-73395c5c1632"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.711722 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "12d9875d-e81f-425b-9821-73395c5c1632" (UID: "12d9875d-e81f-425b-9821-73395c5c1632"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.750931 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.751003 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdjm7\" (UniqueName: \"kubernetes.io/projected/12d9875d-e81f-425b-9821-73395c5c1632-kube-api-access-vdjm7\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.751017 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.751029 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.754979 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "12d9875d-e81f-425b-9821-73395c5c1632" (UID: "12d9875d-e81f-425b-9821-73395c5c1632"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:02 crc kubenswrapper[4806]: I1127 10:38:02.852501 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/12d9875d-e81f-425b-9821-73395c5c1632-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.041204 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.083509 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" event={"ID":"12d9875d-e81f-425b-9821-73395c5c1632","Type":"ContainerDied","Data":"e4b4b0b05d709f8613968c546f65f049a51b2b6bcac26895a40aa526c0ba9fea"} Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.083881 4806 scope.go:117] "RemoveContainer" containerID="b141e7d6e9c2fa98de9d080ba8feb89f941c9f7289c1e1e24a206ce2c62d6d20" Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.084031 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66fbd85b65-pnvmt" Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.095331 4806 generic.go:334] "Generic (PLEG): container finished" podID="734efb9b-df1d-4f4f-ba47-7df13a997e3e" containerID="0402c70a5bcd46173938492a55b5b101205e6e7eb8b36f40c3b17fff688bef1d" exitCode=0 Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.095462 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" event={"ID":"734efb9b-df1d-4f4f-ba47-7df13a997e3e","Type":"ContainerDied","Data":"0402c70a5bcd46173938492a55b5b101205e6e7eb8b36f40c3b17fff688bef1d"} Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.126847 4806 generic.go:334] "Generic (PLEG): container finished" podID="68ebcee0-f362-4953-8775-1add49fb4add" containerID="c0629d030f4cbf04c370623ea92033a5f6772e789f55577b4c894bcbe3680479" exitCode=0 Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.127330 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" event={"ID":"68ebcee0-f362-4953-8775-1add49fb4add","Type":"ContainerDied","Data":"c0629d030f4cbf04c370623ea92033a5f6772e789f55577b4c894bcbe3680479"} Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.291438 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-pnvmt"] Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.309594 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66fbd85b65-pnvmt"] Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.800020 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.989178 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlrgx\" (UniqueName: \"kubernetes.io/projected/734efb9b-df1d-4f4f-ba47-7df13a997e3e-kube-api-access-zlrgx\") pod \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.989313 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-ovsdbserver-sb\") pod \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.989343 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-dns-svc\") pod \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.989406 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-config\") pod \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " Nov 27 10:38:03 crc kubenswrapper[4806]: I1127 10:38:03.989447 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-ovsdbserver-nb\") pod \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\" (UID: \"734efb9b-df1d-4f4f-ba47-7df13a997e3e\") " Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.002883 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/734efb9b-df1d-4f4f-ba47-7df13a997e3e-kube-api-access-zlrgx" (OuterVolumeSpecName: "kube-api-access-zlrgx") pod "734efb9b-df1d-4f4f-ba47-7df13a997e3e" (UID: "734efb9b-df1d-4f4f-ba47-7df13a997e3e"). InnerVolumeSpecName "kube-api-access-zlrgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.020589 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-config" (OuterVolumeSpecName: "config") pod "734efb9b-df1d-4f4f-ba47-7df13a997e3e" (UID: "734efb9b-df1d-4f4f-ba47-7df13a997e3e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.024790 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "734efb9b-df1d-4f4f-ba47-7df13a997e3e" (UID: "734efb9b-df1d-4f4f-ba47-7df13a997e3e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.042312 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "734efb9b-df1d-4f4f-ba47-7df13a997e3e" (UID: "734efb9b-df1d-4f4f-ba47-7df13a997e3e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.050319 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "734efb9b-df1d-4f4f-ba47-7df13a997e3e" (UID: "734efb9b-df1d-4f4f-ba47-7df13a997e3e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.102509 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.102550 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.102560 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.102571 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/734efb9b-df1d-4f4f-ba47-7df13a997e3e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.102583 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlrgx\" (UniqueName: \"kubernetes.io/projected/734efb9b-df1d-4f4f-ba47-7df13a997e3e-kube-api-access-zlrgx\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.136128 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12d9875d-e81f-425b-9821-73395c5c1632" path="/var/lib/kubelet/pods/12d9875d-e81f-425b-9821-73395c5c1632/volumes" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.169497 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" event={"ID":"734efb9b-df1d-4f4f-ba47-7df13a997e3e","Type":"ContainerDied","Data":"1049580d8d8a5a36485318937ec6c97cfc671b49451795ec768c9919c961fa65"} Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.169586 4806 scope.go:117] "RemoveContainer" containerID="0402c70a5bcd46173938492a55b5b101205e6e7eb8b36f40c3b17fff688bef1d" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.169811 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf59f66bf-8mddk" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.177027 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" event={"ID":"68ebcee0-f362-4953-8775-1add49fb4add","Type":"ContainerStarted","Data":"49bf83bfca2471a29892592b546f2f49c3dcc225bc37d76e3e0df21e7b9747d1"} Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.177254 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.230332 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" podStartSLOduration=4.230289891 podStartE2EDuration="4.230289891s" podCreationTimestamp="2025-11-27 10:38:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:38:04.209884424 +0000 UTC m=+988.796475188" watchObservedRunningTime="2025-11-27 10:38:04.230289891 +0000 UTC m=+988.816880655" Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.271249 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-8mddk"] Nov 27 10:38:04 crc kubenswrapper[4806]: I1127 10:38:04.281058 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bf59f66bf-8mddk"] Nov 27 10:38:06 crc kubenswrapper[4806]: I1127 10:38:06.183800 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="734efb9b-df1d-4f4f-ba47-7df13a997e3e" path="/var/lib/kubelet/pods/734efb9b-df1d-4f4f-ba47-7df13a997e3e/volumes" Nov 27 10:38:07 crc kubenswrapper[4806]: I1127 10:38:07.227828 4806 generic.go:334] "Generic (PLEG): container finished" podID="7dad6a31-52cc-4921-b585-2a69c9ebc607" containerID="b5f34a455734a87a2d12b56355e08ede8ebdacaade0e860817dc1ea7ce12d66b" exitCode=0 Nov 27 10:38:07 crc kubenswrapper[4806]: I1127 10:38:07.228223 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-665mv" event={"ID":"7dad6a31-52cc-4921-b585-2a69c9ebc607","Type":"ContainerDied","Data":"b5f34a455734a87a2d12b56355e08ede8ebdacaade0e860817dc1ea7ce12d66b"} Nov 27 10:38:10 crc kubenswrapper[4806]: I1127 10:38:10.719346 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:10 crc kubenswrapper[4806]: I1127 10:38:10.808494 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-9hn6t"] Nov 27 10:38:10 crc kubenswrapper[4806]: I1127 10:38:10.808856 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-9hn6t" podUID="9aec6519-44e3-499c-bdce-151bf8a13a25" containerName="dnsmasq-dns" containerID="cri-o://897eb0bb82fed8ef6c3e6470f0373b2a88c89dc8027e91bf6b052f5bc063db6a" gracePeriod=10 Nov 27 10:38:11 crc kubenswrapper[4806]: I1127 10:38:11.278924 4806 generic.go:334] "Generic (PLEG): container finished" podID="9aec6519-44e3-499c-bdce-151bf8a13a25" containerID="897eb0bb82fed8ef6c3e6470f0373b2a88c89dc8027e91bf6b052f5bc063db6a" exitCode=0 Nov 27 10:38:11 crc kubenswrapper[4806]: I1127 10:38:11.278994 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-9hn6t" event={"ID":"9aec6519-44e3-499c-bdce-151bf8a13a25","Type":"ContainerDied","Data":"897eb0bb82fed8ef6c3e6470f0373b2a88c89dc8027e91bf6b052f5bc063db6a"} Nov 27 10:38:11 crc kubenswrapper[4806]: I1127 10:38:11.685653 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-9hn6t" podUID="9aec6519-44e3-499c-bdce-151bf8a13a25" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.108:5353: connect: connection refused" Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.642336 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-665mv" Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.718328 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sfk7\" (UniqueName: \"kubernetes.io/projected/7dad6a31-52cc-4921-b585-2a69c9ebc607-kube-api-access-4sfk7\") pod \"7dad6a31-52cc-4921-b585-2a69c9ebc607\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.718452 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-fernet-keys\") pod \"7dad6a31-52cc-4921-b585-2a69c9ebc607\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.718505 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-scripts\") pod \"7dad6a31-52cc-4921-b585-2a69c9ebc607\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.718624 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-config-data\") pod \"7dad6a31-52cc-4921-b585-2a69c9ebc607\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.718738 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-combined-ca-bundle\") pod \"7dad6a31-52cc-4921-b585-2a69c9ebc607\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.718774 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-credential-keys\") pod \"7dad6a31-52cc-4921-b585-2a69c9ebc607\" (UID: \"7dad6a31-52cc-4921-b585-2a69c9ebc607\") " Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.726677 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7dad6a31-52cc-4921-b585-2a69c9ebc607" (UID: "7dad6a31-52cc-4921-b585-2a69c9ebc607"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.726892 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dad6a31-52cc-4921-b585-2a69c9ebc607-kube-api-access-4sfk7" (OuterVolumeSpecName: "kube-api-access-4sfk7") pod "7dad6a31-52cc-4921-b585-2a69c9ebc607" (UID: "7dad6a31-52cc-4921-b585-2a69c9ebc607"). InnerVolumeSpecName "kube-api-access-4sfk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.731979 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-scripts" (OuterVolumeSpecName: "scripts") pod "7dad6a31-52cc-4921-b585-2a69c9ebc607" (UID: "7dad6a31-52cc-4921-b585-2a69c9ebc607"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.737951 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "7dad6a31-52cc-4921-b585-2a69c9ebc607" (UID: "7dad6a31-52cc-4921-b585-2a69c9ebc607"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.754705 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7dad6a31-52cc-4921-b585-2a69c9ebc607" (UID: "7dad6a31-52cc-4921-b585-2a69c9ebc607"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.761547 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-config-data" (OuterVolumeSpecName: "config-data") pod "7dad6a31-52cc-4921-b585-2a69c9ebc607" (UID: "7dad6a31-52cc-4921-b585-2a69c9ebc607"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.821607 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.821654 4806 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.821664 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sfk7\" (UniqueName: \"kubernetes.io/projected/7dad6a31-52cc-4921-b585-2a69c9ebc607-kube-api-access-4sfk7\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.821678 4806 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.821688 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:12 crc kubenswrapper[4806]: I1127 10:38:12.821696 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dad6a31-52cc-4921-b585-2a69c9ebc607-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.306213 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9q6vl"] Nov 27 10:38:13 crc kubenswrapper[4806]: E1127 10:38:13.306963 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d9875d-e81f-425b-9821-73395c5c1632" containerName="init" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.306981 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d9875d-e81f-425b-9821-73395c5c1632" containerName="init" Nov 27 10:38:13 crc kubenswrapper[4806]: E1127 10:38:13.307024 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="734efb9b-df1d-4f4f-ba47-7df13a997e3e" containerName="init" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.307031 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="734efb9b-df1d-4f4f-ba47-7df13a997e3e" containerName="init" Nov 27 10:38:13 crc kubenswrapper[4806]: E1127 10:38:13.307051 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dad6a31-52cc-4921-b585-2a69c9ebc607" containerName="keystone-bootstrap" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.307059 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dad6a31-52cc-4921-b585-2a69c9ebc607" containerName="keystone-bootstrap" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.307443 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dad6a31-52cc-4921-b585-2a69c9ebc607" containerName="keystone-bootstrap" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.307469 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="734efb9b-df1d-4f4f-ba47-7df13a997e3e" containerName="init" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.307490 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="12d9875d-e81f-425b-9821-73395c5c1632" containerName="init" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.310195 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.322089 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9q6vl"] Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.336909 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-665mv" event={"ID":"7dad6a31-52cc-4921-b585-2a69c9ebc607","Type":"ContainerDied","Data":"e9b58bdd1816684d4ec0778b4ba938d63bf26ea7545af2ef90f7f7e43ab3aa91"} Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.336973 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9b58bdd1816684d4ec0778b4ba938d63bf26ea7545af2ef90f7f7e43ab3aa91" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.337310 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-665mv" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.341775 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b50a58-b14e-4ae3-ae29-98e5181aa828-utilities\") pod \"redhat-operators-9q6vl\" (UID: \"95b50a58-b14e-4ae3-ae29-98e5181aa828\") " pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.341865 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b50a58-b14e-4ae3-ae29-98e5181aa828-catalog-content\") pod \"redhat-operators-9q6vl\" (UID: \"95b50a58-b14e-4ae3-ae29-98e5181aa828\") " pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.341956 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvjnz\" (UniqueName: \"kubernetes.io/projected/95b50a58-b14e-4ae3-ae29-98e5181aa828-kube-api-access-pvjnz\") pod \"redhat-operators-9q6vl\" (UID: \"95b50a58-b14e-4ae3-ae29-98e5181aa828\") " pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.444375 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b50a58-b14e-4ae3-ae29-98e5181aa828-catalog-content\") pod \"redhat-operators-9q6vl\" (UID: \"95b50a58-b14e-4ae3-ae29-98e5181aa828\") " pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.444665 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvjnz\" (UniqueName: \"kubernetes.io/projected/95b50a58-b14e-4ae3-ae29-98e5181aa828-kube-api-access-pvjnz\") pod \"redhat-operators-9q6vl\" (UID: \"95b50a58-b14e-4ae3-ae29-98e5181aa828\") " pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.444844 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b50a58-b14e-4ae3-ae29-98e5181aa828-utilities\") pod \"redhat-operators-9q6vl\" (UID: \"95b50a58-b14e-4ae3-ae29-98e5181aa828\") " pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.444928 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b50a58-b14e-4ae3-ae29-98e5181aa828-catalog-content\") pod \"redhat-operators-9q6vl\" (UID: \"95b50a58-b14e-4ae3-ae29-98e5181aa828\") " pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.445364 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b50a58-b14e-4ae3-ae29-98e5181aa828-utilities\") pod \"redhat-operators-9q6vl\" (UID: \"95b50a58-b14e-4ae3-ae29-98e5181aa828\") " pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.462367 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvjnz\" (UniqueName: \"kubernetes.io/projected/95b50a58-b14e-4ae3-ae29-98e5181aa828-kube-api-access-pvjnz\") pod \"redhat-operators-9q6vl\" (UID: \"95b50a58-b14e-4ae3-ae29-98e5181aa828\") " pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.649892 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.845645 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-665mv"] Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.852259 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-665mv"] Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.945072 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jkq5l"] Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.949164 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.951927 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.952144 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.952402 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.952610 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.954865 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxgbr" Nov 27 10:38:13 crc kubenswrapper[4806]: I1127 10:38:13.956870 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jkq5l"] Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.056090 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-fernet-keys\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.056529 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-config-data\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.056668 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-combined-ca-bundle\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.056813 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-credential-keys\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.056979 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-scripts\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.057137 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz97g\" (UniqueName: \"kubernetes.io/projected/c109a218-41c4-4872-9fee-fbba98bb0143-kube-api-access-dz97g\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.132828 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dad6a31-52cc-4921-b585-2a69c9ebc607" path="/var/lib/kubelet/pods/7dad6a31-52cc-4921-b585-2a69c9ebc607/volumes" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.159468 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-scripts\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.159522 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz97g\" (UniqueName: \"kubernetes.io/projected/c109a218-41c4-4872-9fee-fbba98bb0143-kube-api-access-dz97g\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.159620 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-fernet-keys\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.159664 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-config-data\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.159686 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-combined-ca-bundle\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.159719 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-credential-keys\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.170129 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-config-data\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.170977 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-fernet-keys\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.171701 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-combined-ca-bundle\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.210518 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-scripts\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.210826 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-credential-keys\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.242044 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz97g\" (UniqueName: \"kubernetes.io/projected/c109a218-41c4-4872-9fee-fbba98bb0143-kube-api-access-dz97g\") pod \"keystone-bootstrap-jkq5l\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:14 crc kubenswrapper[4806]: I1127 10:38:14.268808 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:15 crc kubenswrapper[4806]: E1127 10:38:15.396440 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Nov 27 10:38:15 crc kubenswrapper[4806]: E1127 10:38:15.396970 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zpgc2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-hz79k_openstack(b289081e-bb54-4046-9e62-37e47cacca99): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:38:15 crc kubenswrapper[4806]: E1127 10:38:15.398220 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-hz79k" podUID="b289081e-bb54-4046-9e62-37e47cacca99" Nov 27 10:38:16 crc kubenswrapper[4806]: E1127 10:38:16.370554 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-hz79k" podUID="b289081e-bb54-4046-9e62-37e47cacca99" Nov 27 10:38:21 crc kubenswrapper[4806]: I1127 10:38:21.686459 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-9hn6t" podUID="9aec6519-44e3-499c-bdce-151bf8a13a25" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.108:5353: i/o timeout" Nov 27 10:38:25 crc kubenswrapper[4806]: E1127 10:38:25.737178 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 27 10:38:25 crc kubenswrapper[4806]: E1127 10:38:25.737928 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lnt7d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-cwqkj_openstack(552d400b-d5b2-40c8-9654-bfc70ab5bab6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:38:25 crc kubenswrapper[4806]: E1127 10:38:25.740303 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-cwqkj" podUID="552d400b-d5b2-40c8-9654-bfc70ab5bab6" Nov 27 10:38:26 crc kubenswrapper[4806]: E1127 10:38:26.191890 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 27 10:38:26 crc kubenswrapper[4806]: E1127 10:38:26.192534 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pzbjf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-9vkqx_openstack(11b83e17-71ca-4f23-ad17-950a5de85a12): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 10:38:26 crc kubenswrapper[4806]: E1127 10:38:26.193994 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-9vkqx" podUID="11b83e17-71ca-4f23-ad17-950a5de85a12" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.407779 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.461335 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-9hn6t" event={"ID":"9aec6519-44e3-499c-bdce-151bf8a13a25","Type":"ContainerDied","Data":"fe4cf7cb01cdfec80a130a332a5b62fe1d63bcbfaebe70bdfb79fa5bc015f72e"} Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.461393 4806 scope.go:117] "RemoveContainer" containerID="897eb0bb82fed8ef6c3e6470f0373b2a88c89dc8027e91bf6b052f5bc063db6a" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.461567 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:38:26 crc kubenswrapper[4806]: E1127 10:38:26.463553 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-9vkqx" podUID="11b83e17-71ca-4f23-ad17-950a5de85a12" Nov 27 10:38:26 crc kubenswrapper[4806]: E1127 10:38:26.463752 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-cwqkj" podUID="552d400b-d5b2-40c8-9654-bfc70ab5bab6" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.504630 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-ovsdbserver-nb\") pod \"9aec6519-44e3-499c-bdce-151bf8a13a25\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.504702 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-ovsdbserver-sb\") pod \"9aec6519-44e3-499c-bdce-151bf8a13a25\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.504853 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-dns-svc\") pod \"9aec6519-44e3-499c-bdce-151bf8a13a25\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.504876 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrncd\" (UniqueName: \"kubernetes.io/projected/9aec6519-44e3-499c-bdce-151bf8a13a25-kube-api-access-qrncd\") pod \"9aec6519-44e3-499c-bdce-151bf8a13a25\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.505192 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-config\") pod \"9aec6519-44e3-499c-bdce-151bf8a13a25\" (UID: \"9aec6519-44e3-499c-bdce-151bf8a13a25\") " Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.523757 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aec6519-44e3-499c-bdce-151bf8a13a25-kube-api-access-qrncd" (OuterVolumeSpecName: "kube-api-access-qrncd") pod "9aec6519-44e3-499c-bdce-151bf8a13a25" (UID: "9aec6519-44e3-499c-bdce-151bf8a13a25"). InnerVolumeSpecName "kube-api-access-qrncd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.525818 4806 scope.go:117] "RemoveContainer" containerID="953b0dc7f1e2e636ab4ac6979a43ca6c14266fc05e6a46a0f7bfaaf994379a72" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.559772 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9aec6519-44e3-499c-bdce-151bf8a13a25" (UID: "9aec6519-44e3-499c-bdce-151bf8a13a25"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.575140 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-config" (OuterVolumeSpecName: "config") pod "9aec6519-44e3-499c-bdce-151bf8a13a25" (UID: "9aec6519-44e3-499c-bdce-151bf8a13a25"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.581354 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9aec6519-44e3-499c-bdce-151bf8a13a25" (UID: "9aec6519-44e3-499c-bdce-151bf8a13a25"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.594792 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9aec6519-44e3-499c-bdce-151bf8a13a25" (UID: "9aec6519-44e3-499c-bdce-151bf8a13a25"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.608244 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.608281 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.608298 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.608325 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrncd\" (UniqueName: \"kubernetes.io/projected/9aec6519-44e3-499c-bdce-151bf8a13a25-kube-api-access-qrncd\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.608336 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aec6519-44e3-499c-bdce-151bf8a13a25-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.687125 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-9hn6t" podUID="9aec6519-44e3-499c-bdce-151bf8a13a25" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.108:5353: i/o timeout" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.687454 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-9hn6t" Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.697305 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jkq5l"] Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.782710 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9q6vl"] Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.809406 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-9hn6t"] Nov 27 10:38:26 crc kubenswrapper[4806]: I1127 10:38:26.821424 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-9hn6t"] Nov 27 10:38:27 crc kubenswrapper[4806]: I1127 10:38:27.469836 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05ff370-09c4-45c1-b101-a6b277a34d3d","Type":"ContainerStarted","Data":"07377df1241e43c09f0052ed513662c35a9c3b54c8608fdc9491c4e98408cdf1"} Nov 27 10:38:27 crc kubenswrapper[4806]: I1127 10:38:27.471372 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jkq5l" event={"ID":"c109a218-41c4-4872-9fee-fbba98bb0143","Type":"ContainerStarted","Data":"40b85bffaf47fe48ed4844f796570b481a16e94be64512d19ce2624d90183d17"} Nov 27 10:38:27 crc kubenswrapper[4806]: I1127 10:38:27.471434 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jkq5l" event={"ID":"c109a218-41c4-4872-9fee-fbba98bb0143","Type":"ContainerStarted","Data":"25fe750f0a6a0a1d32e2c847be4a89e8cb88643d8352dc85333bea77acf10d91"} Nov 27 10:38:27 crc kubenswrapper[4806]: I1127 10:38:27.473939 4806 generic.go:334] "Generic (PLEG): container finished" podID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerID="9eab18251d6748b7bdfacf0f098d4a00a5a39e3890097d25103566c1e599117f" exitCode=0 Nov 27 10:38:27 crc kubenswrapper[4806]: I1127 10:38:27.474044 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9q6vl" event={"ID":"95b50a58-b14e-4ae3-ae29-98e5181aa828","Type":"ContainerDied","Data":"9eab18251d6748b7bdfacf0f098d4a00a5a39e3890097d25103566c1e599117f"} Nov 27 10:38:27 crc kubenswrapper[4806]: I1127 10:38:27.474096 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9q6vl" event={"ID":"95b50a58-b14e-4ae3-ae29-98e5181aa828","Type":"ContainerStarted","Data":"522117fb9f162931de795793793aa168b4ac8f138689e77fedf4291c96570e3f"} Nov 27 10:38:27 crc kubenswrapper[4806]: I1127 10:38:27.476468 4806 generic.go:334] "Generic (PLEG): container finished" podID="94086dd9-9626-49a4-ad72-af183dc0a161" containerID="d2a5eda0802027071d1e68bf8b7f9de6516a5df948f774197491684a44f05028" exitCode=0 Nov 27 10:38:27 crc kubenswrapper[4806]: I1127 10:38:27.476548 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-h7fbq" event={"ID":"94086dd9-9626-49a4-ad72-af183dc0a161","Type":"ContainerDied","Data":"d2a5eda0802027071d1e68bf8b7f9de6516a5df948f774197491684a44f05028"} Nov 27 10:38:27 crc kubenswrapper[4806]: I1127 10:38:27.499254 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jkq5l" podStartSLOduration=14.499219125 podStartE2EDuration="14.499219125s" podCreationTimestamp="2025-11-27 10:38:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:38:27.488925209 +0000 UTC m=+1012.075515973" watchObservedRunningTime="2025-11-27 10:38:27.499219125 +0000 UTC m=+1012.085809889" Nov 27 10:38:28 crc kubenswrapper[4806]: I1127 10:38:28.127056 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aec6519-44e3-499c-bdce-151bf8a13a25" path="/var/lib/kubelet/pods/9aec6519-44e3-499c-bdce-151bf8a13a25/volumes" Nov 27 10:38:28 crc kubenswrapper[4806]: I1127 10:38:28.501359 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9q6vl" event={"ID":"95b50a58-b14e-4ae3-ae29-98e5181aa828","Type":"ContainerStarted","Data":"b8b296b8d2494494a350a87831356c852babf046758d6744b025bcd0010c9fa2"} Nov 27 10:38:28 crc kubenswrapper[4806]: I1127 10:38:28.504716 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05ff370-09c4-45c1-b101-a6b277a34d3d","Type":"ContainerStarted","Data":"9c03a47efe11a0973c4aaa2985a4039e32eb53c1fe4d3d8393c24291624b7bfb"} Nov 27 10:38:28 crc kubenswrapper[4806]: I1127 10:38:28.916664 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-h7fbq" Nov 27 10:38:28 crc kubenswrapper[4806]: I1127 10:38:28.952094 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94086dd9-9626-49a4-ad72-af183dc0a161-combined-ca-bundle\") pod \"94086dd9-9626-49a4-ad72-af183dc0a161\" (UID: \"94086dd9-9626-49a4-ad72-af183dc0a161\") " Nov 27 10:38:28 crc kubenswrapper[4806]: I1127 10:38:28.952352 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/94086dd9-9626-49a4-ad72-af183dc0a161-config\") pod \"94086dd9-9626-49a4-ad72-af183dc0a161\" (UID: \"94086dd9-9626-49a4-ad72-af183dc0a161\") " Nov 27 10:38:28 crc kubenswrapper[4806]: I1127 10:38:28.952380 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kv2qr\" (UniqueName: \"kubernetes.io/projected/94086dd9-9626-49a4-ad72-af183dc0a161-kube-api-access-kv2qr\") pod \"94086dd9-9626-49a4-ad72-af183dc0a161\" (UID: \"94086dd9-9626-49a4-ad72-af183dc0a161\") " Nov 27 10:38:28 crc kubenswrapper[4806]: I1127 10:38:28.958868 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94086dd9-9626-49a4-ad72-af183dc0a161-kube-api-access-kv2qr" (OuterVolumeSpecName: "kube-api-access-kv2qr") pod "94086dd9-9626-49a4-ad72-af183dc0a161" (UID: "94086dd9-9626-49a4-ad72-af183dc0a161"). InnerVolumeSpecName "kube-api-access-kv2qr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.000927 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94086dd9-9626-49a4-ad72-af183dc0a161-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94086dd9-9626-49a4-ad72-af183dc0a161" (UID: "94086dd9-9626-49a4-ad72-af183dc0a161"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.017465 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94086dd9-9626-49a4-ad72-af183dc0a161-config" (OuterVolumeSpecName: "config") pod "94086dd9-9626-49a4-ad72-af183dc0a161" (UID: "94086dd9-9626-49a4-ad72-af183dc0a161"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.055222 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/94086dd9-9626-49a4-ad72-af183dc0a161-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.055274 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kv2qr\" (UniqueName: \"kubernetes.io/projected/94086dd9-9626-49a4-ad72-af183dc0a161-kube-api-access-kv2qr\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.055285 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94086dd9-9626-49a4-ad72-af183dc0a161-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.525773 4806 generic.go:334] "Generic (PLEG): container finished" podID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerID="b8b296b8d2494494a350a87831356c852babf046758d6744b025bcd0010c9fa2" exitCode=0 Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.525861 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9q6vl" event={"ID":"95b50a58-b14e-4ae3-ae29-98e5181aa828","Type":"ContainerDied","Data":"b8b296b8d2494494a350a87831356c852babf046758d6744b025bcd0010c9fa2"} Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.528978 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-h7fbq" event={"ID":"94086dd9-9626-49a4-ad72-af183dc0a161","Type":"ContainerDied","Data":"6a43ed6818a88daf6385e875eae5e84d01e9f8a360ee97893b3358ffb712478c"} Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.529004 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a43ed6818a88daf6385e875eae5e84d01e9f8a360ee97893b3358ffb712478c" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.537965 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-h7fbq" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.732710 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-vpvdm"] Nov 27 10:38:29 crc kubenswrapper[4806]: E1127 10:38:29.735295 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94086dd9-9626-49a4-ad72-af183dc0a161" containerName="neutron-db-sync" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.738903 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="94086dd9-9626-49a4-ad72-af183dc0a161" containerName="neutron-db-sync" Nov 27 10:38:29 crc kubenswrapper[4806]: E1127 10:38:29.739123 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aec6519-44e3-499c-bdce-151bf8a13a25" containerName="init" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.739205 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aec6519-44e3-499c-bdce-151bf8a13a25" containerName="init" Nov 27 10:38:29 crc kubenswrapper[4806]: E1127 10:38:29.739301 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aec6519-44e3-499c-bdce-151bf8a13a25" containerName="dnsmasq-dns" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.739368 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aec6519-44e3-499c-bdce-151bf8a13a25" containerName="dnsmasq-dns" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.739734 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="94086dd9-9626-49a4-ad72-af183dc0a161" containerName="neutron-db-sync" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.739817 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aec6519-44e3-499c-bdce-151bf8a13a25" containerName="dnsmasq-dns" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.741554 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.750152 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-vpvdm"] Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.766704 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-config\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.766926 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.766999 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx4c4\" (UniqueName: \"kubernetes.io/projected/95cec57d-0ffa-497f-9850-b460bb6a2b53-kube-api-access-hx4c4\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.767096 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.767215 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.870023 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-config\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.870099 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.870118 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx4c4\" (UniqueName: \"kubernetes.io/projected/95cec57d-0ffa-497f-9850-b460bb6a2b53-kube-api-access-hx4c4\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.870157 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.870178 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.870984 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.871856 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.871892 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-config\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.872653 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.902843 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-88c6547cb-ngwks"] Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.904104 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.904626 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx4c4\" (UniqueName: \"kubernetes.io/projected/95cec57d-0ffa-497f-9850-b460bb6a2b53-kube-api-access-hx4c4\") pod \"dnsmasq-dns-5f66db59b9-vpvdm\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.920521 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.920811 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-9v77k" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.921163 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.921264 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 27 10:38:29 crc kubenswrapper[4806]: I1127 10:38:29.933126 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-88c6547cb-ngwks"] Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.067510 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.072513 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-combined-ca-bundle\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.072659 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-ovndb-tls-certs\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.072758 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-config\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.072796 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zms72\" (UniqueName: \"kubernetes.io/projected/c5760001-0bdf-40bb-86ab-34017e1c5297-kube-api-access-zms72\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.072927 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-httpd-config\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.174581 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-ovndb-tls-certs\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.174650 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-config\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.174676 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zms72\" (UniqueName: \"kubernetes.io/projected/c5760001-0bdf-40bb-86ab-34017e1c5297-kube-api-access-zms72\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.174720 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-httpd-config\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.174807 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-combined-ca-bundle\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.179007 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-combined-ca-bundle\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.195314 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-config\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.195444 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-httpd-config\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.202809 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-ovndb-tls-certs\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.213570 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zms72\" (UniqueName: \"kubernetes.io/projected/c5760001-0bdf-40bb-86ab-34017e1c5297-kube-api-access-zms72\") pod \"neutron-88c6547cb-ngwks\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.253975 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:30 crc kubenswrapper[4806]: I1127 10:38:30.730262 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-vpvdm"] Nov 27 10:38:31 crc kubenswrapper[4806]: I1127 10:38:31.129325 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-88c6547cb-ngwks"] Nov 27 10:38:31 crc kubenswrapper[4806]: I1127 10:38:31.589377 4806 generic.go:334] "Generic (PLEG): container finished" podID="95cec57d-0ffa-497f-9850-b460bb6a2b53" containerID="975e1efecc3fc6a8a63c241af367b867208fbb71c479632d8d68dca4dac67a79" exitCode=0 Nov 27 10:38:31 crc kubenswrapper[4806]: I1127 10:38:31.589446 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" event={"ID":"95cec57d-0ffa-497f-9850-b460bb6a2b53","Type":"ContainerDied","Data":"975e1efecc3fc6a8a63c241af367b867208fbb71c479632d8d68dca4dac67a79"} Nov 27 10:38:31 crc kubenswrapper[4806]: I1127 10:38:31.589481 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" event={"ID":"95cec57d-0ffa-497f-9850-b460bb6a2b53","Type":"ContainerStarted","Data":"b719a5c93722e266da4ebe243f4e6f9a11456d9b1d84822f328af6670ee50574"} Nov 27 10:38:31 crc kubenswrapper[4806]: I1127 10:38:31.592435 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-88c6547cb-ngwks" event={"ID":"c5760001-0bdf-40bb-86ab-34017e1c5297","Type":"ContainerStarted","Data":"3a308f295b808edae84b73c41ec07fe45394d50412d4c94848ea3c1c0df9124e"} Nov 27 10:38:31 crc kubenswrapper[4806]: I1127 10:38:31.592507 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-88c6547cb-ngwks" event={"ID":"c5760001-0bdf-40bb-86ab-34017e1c5297","Type":"ContainerStarted","Data":"9dd285faca5c35f034597e2edf7906617b876286ddc2744e6bdeff2986f64781"} Nov 27 10:38:31 crc kubenswrapper[4806]: I1127 10:38:31.601270 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hz79k" event={"ID":"b289081e-bb54-4046-9e62-37e47cacca99","Type":"ContainerStarted","Data":"884656ef95050d07810144fe740ae52ba53cbdf5fa5c5756fdba61f8c93ed143"} Nov 27 10:38:31 crc kubenswrapper[4806]: I1127 10:38:31.652709 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-hz79k" podStartSLOduration=4.271428429 podStartE2EDuration="32.652689945s" podCreationTimestamp="2025-11-27 10:37:59 +0000 UTC" firstStartedPulling="2025-11-27 10:38:01.815005712 +0000 UTC m=+986.401596476" lastFinishedPulling="2025-11-27 10:38:30.196267228 +0000 UTC m=+1014.782857992" observedRunningTime="2025-11-27 10:38:31.64710152 +0000 UTC m=+1016.233692294" watchObservedRunningTime="2025-11-27 10:38:31.652689945 +0000 UTC m=+1016.239280709" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.390929 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5c76df9f57-fglss"] Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.394120 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.397078 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.397506 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.428286 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c76df9f57-fglss"] Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.559989 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-config\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.560075 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-httpd-config\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.560107 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-internal-tls-certs\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.560134 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-public-tls-certs\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.560988 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-combined-ca-bundle\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.561019 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz6kc\" (UniqueName: \"kubernetes.io/projected/93660245-72f2-4792-af51-3ac612926729-kube-api-access-nz6kc\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.561051 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-ovndb-tls-certs\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.627714 4806 generic.go:334] "Generic (PLEG): container finished" podID="c109a218-41c4-4872-9fee-fbba98bb0143" containerID="40b85bffaf47fe48ed4844f796570b481a16e94be64512d19ce2624d90183d17" exitCode=0 Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.627780 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jkq5l" event={"ID":"c109a218-41c4-4872-9fee-fbba98bb0143","Type":"ContainerDied","Data":"40b85bffaf47fe48ed4844f796570b481a16e94be64512d19ce2624d90183d17"} Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.662696 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-config\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.662810 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-httpd-config\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.662855 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-internal-tls-certs\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.662887 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-public-tls-certs\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.662926 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-combined-ca-bundle\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.662953 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz6kc\" (UniqueName: \"kubernetes.io/projected/93660245-72f2-4792-af51-3ac612926729-kube-api-access-nz6kc\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.663010 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-ovndb-tls-certs\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.672118 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-public-tls-certs\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.673509 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-internal-tls-certs\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.676940 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-ovndb-tls-certs\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.682347 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-combined-ca-bundle\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.685165 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-httpd-config\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.694958 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/93660245-72f2-4792-af51-3ac612926729-config\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.697094 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz6kc\" (UniqueName: \"kubernetes.io/projected/93660245-72f2-4792-af51-3ac612926729-kube-api-access-nz6kc\") pod \"neutron-5c76df9f57-fglss\" (UID: \"93660245-72f2-4792-af51-3ac612926729\") " pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:33 crc kubenswrapper[4806]: I1127 10:38:33.715626 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:34 crc kubenswrapper[4806]: I1127 10:38:34.645306 4806 generic.go:334] "Generic (PLEG): container finished" podID="b289081e-bb54-4046-9e62-37e47cacca99" containerID="884656ef95050d07810144fe740ae52ba53cbdf5fa5c5756fdba61f8c93ed143" exitCode=0 Nov 27 10:38:34 crc kubenswrapper[4806]: I1127 10:38:34.645361 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hz79k" event={"ID":"b289081e-bb54-4046-9e62-37e47cacca99","Type":"ContainerDied","Data":"884656ef95050d07810144fe740ae52ba53cbdf5fa5c5756fdba61f8c93ed143"} Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.035077 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.093918 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hz79k" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.147141 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-scripts\") pod \"c109a218-41c4-4872-9fee-fbba98bb0143\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.147647 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dz97g\" (UniqueName: \"kubernetes.io/projected/c109a218-41c4-4872-9fee-fbba98bb0143-kube-api-access-dz97g\") pod \"c109a218-41c4-4872-9fee-fbba98bb0143\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.147747 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-combined-ca-bundle\") pod \"c109a218-41c4-4872-9fee-fbba98bb0143\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.147972 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-credential-keys\") pod \"c109a218-41c4-4872-9fee-fbba98bb0143\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.148133 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-fernet-keys\") pod \"c109a218-41c4-4872-9fee-fbba98bb0143\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.153811 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-config-data\") pod \"c109a218-41c4-4872-9fee-fbba98bb0143\" (UID: \"c109a218-41c4-4872-9fee-fbba98bb0143\") " Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.168910 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c109a218-41c4-4872-9fee-fbba98bb0143" (UID: "c109a218-41c4-4872-9fee-fbba98bb0143"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.170916 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-scripts" (OuterVolumeSpecName: "scripts") pod "c109a218-41c4-4872-9fee-fbba98bb0143" (UID: "c109a218-41c4-4872-9fee-fbba98bb0143"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.177376 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c109a218-41c4-4872-9fee-fbba98bb0143" (UID: "c109a218-41c4-4872-9fee-fbba98bb0143"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.186159 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c109a218-41c4-4872-9fee-fbba98bb0143-kube-api-access-dz97g" (OuterVolumeSpecName: "kube-api-access-dz97g") pod "c109a218-41c4-4872-9fee-fbba98bb0143" (UID: "c109a218-41c4-4872-9fee-fbba98bb0143"). InnerVolumeSpecName "kube-api-access-dz97g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.257410 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-combined-ca-bundle\") pod \"b289081e-bb54-4046-9e62-37e47cacca99\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.257461 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b289081e-bb54-4046-9e62-37e47cacca99-logs\") pod \"b289081e-bb54-4046-9e62-37e47cacca99\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.257484 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-scripts\") pod \"b289081e-bb54-4046-9e62-37e47cacca99\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.257563 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpgc2\" (UniqueName: \"kubernetes.io/projected/b289081e-bb54-4046-9e62-37e47cacca99-kube-api-access-zpgc2\") pod \"b289081e-bb54-4046-9e62-37e47cacca99\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.257611 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-config-data\") pod \"b289081e-bb54-4046-9e62-37e47cacca99\" (UID: \"b289081e-bb54-4046-9e62-37e47cacca99\") " Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.257943 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b289081e-bb54-4046-9e62-37e47cacca99-logs" (OuterVolumeSpecName: "logs") pod "b289081e-bb54-4046-9e62-37e47cacca99" (UID: "b289081e-bb54-4046-9e62-37e47cacca99"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.258063 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.258079 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dz97g\" (UniqueName: \"kubernetes.io/projected/c109a218-41c4-4872-9fee-fbba98bb0143-kube-api-access-dz97g\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.258090 4806 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.258101 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b289081e-bb54-4046-9e62-37e47cacca99-logs\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.258109 4806 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.274680 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-scripts" (OuterVolumeSpecName: "scripts") pod "b289081e-bb54-4046-9e62-37e47cacca99" (UID: "b289081e-bb54-4046-9e62-37e47cacca99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.284933 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b289081e-bb54-4046-9e62-37e47cacca99-kube-api-access-zpgc2" (OuterVolumeSpecName: "kube-api-access-zpgc2") pod "b289081e-bb54-4046-9e62-37e47cacca99" (UID: "b289081e-bb54-4046-9e62-37e47cacca99"). InnerVolumeSpecName "kube-api-access-zpgc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.315824 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-config-data" (OuterVolumeSpecName: "config-data") pod "c109a218-41c4-4872-9fee-fbba98bb0143" (UID: "c109a218-41c4-4872-9fee-fbba98bb0143"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.327047 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c109a218-41c4-4872-9fee-fbba98bb0143" (UID: "c109a218-41c4-4872-9fee-fbba98bb0143"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.352450 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-config-data" (OuterVolumeSpecName: "config-data") pod "b289081e-bb54-4046-9e62-37e47cacca99" (UID: "b289081e-bb54-4046-9e62-37e47cacca99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.353034 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b289081e-bb54-4046-9e62-37e47cacca99" (UID: "b289081e-bb54-4046-9e62-37e47cacca99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.359139 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.359161 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.359171 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.359182 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpgc2\" (UniqueName: \"kubernetes.io/projected/b289081e-bb54-4046-9e62-37e47cacca99-kube-api-access-zpgc2\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.359192 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b289081e-bb54-4046-9e62-37e47cacca99-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.359202 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c109a218-41c4-4872-9fee-fbba98bb0143-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.519803 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c76df9f57-fglss"] Nov 27 10:38:37 crc kubenswrapper[4806]: W1127 10:38:37.525744 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93660245_72f2_4792_af51_3ac612926729.slice/crio-4a5e518b3f195b7786343460181c8a129028151682b50219b9e574a7b5d1d759 WatchSource:0}: Error finding container 4a5e518b3f195b7786343460181c8a129028151682b50219b9e574a7b5d1d759: Status 404 returned error can't find the container with id 4a5e518b3f195b7786343460181c8a129028151682b50219b9e574a7b5d1d759 Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.677704 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c76df9f57-fglss" event={"ID":"93660245-72f2-4792-af51-3ac612926729","Type":"ContainerStarted","Data":"4a5e518b3f195b7786343460181c8a129028151682b50219b9e574a7b5d1d759"} Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.683148 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05ff370-09c4-45c1-b101-a6b277a34d3d","Type":"ContainerStarted","Data":"183d9372b53f6427808ac6d0ce282b15607ed7035e79538be935f9fddf1eb346"} Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.686059 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" event={"ID":"95cec57d-0ffa-497f-9850-b460bb6a2b53","Type":"ContainerStarted","Data":"3d49bed70224c73e49e5e3ec59a8006b5eb44a25bc9d05864c7f6d01c7acc41c"} Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.686291 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.688827 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-88c6547cb-ngwks" event={"ID":"c5760001-0bdf-40bb-86ab-34017e1c5297","Type":"ContainerStarted","Data":"5daf51ee684219f2b33d04b0532759ea8154cf5898ed727c7a505b432ce17808"} Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.703449 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.709844 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jkq5l" event={"ID":"c109a218-41c4-4872-9fee-fbba98bb0143","Type":"ContainerDied","Data":"25fe750f0a6a0a1d32e2c847be4a89e8cb88643d8352dc85333bea77acf10d91"} Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.709889 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25fe750f0a6a0a1d32e2c847be4a89e8cb88643d8352dc85333bea77acf10d91" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.709987 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jkq5l" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.720505 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hz79k" event={"ID":"b289081e-bb54-4046-9e62-37e47cacca99","Type":"ContainerDied","Data":"e5ea685ebd5fd28b91208d64aef30c5681931774720f7cb0043f46d8afcb6998"} Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.720846 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5ea685ebd5fd28b91208d64aef30c5681931774720f7cb0043f46d8afcb6998" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.721281 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hz79k" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.756758 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9q6vl" event={"ID":"95b50a58-b14e-4ae3-ae29-98e5181aa828","Type":"ContainerStarted","Data":"b3cd2c17c4874ce0e2745f24437a88a7f004d53c35a7ef2f510e9466cb07dd1a"} Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.761997 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" podStartSLOduration=8.76194627 podStartE2EDuration="8.76194627s" podCreationTimestamp="2025-11-27 10:38:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:38:37.72946063 +0000 UTC m=+1022.316051404" watchObservedRunningTime="2025-11-27 10:38:37.76194627 +0000 UTC m=+1022.348537034" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.796221 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-88c6547cb-ngwks" podStartSLOduration=8.79619979 podStartE2EDuration="8.79619979s" podCreationTimestamp="2025-11-27 10:38:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:38:37.775940289 +0000 UTC m=+1022.362531053" watchObservedRunningTime="2025-11-27 10:38:37.79619979 +0000 UTC m=+1022.382790554" Nov 27 10:38:37 crc kubenswrapper[4806]: I1127 10:38:37.821045 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9q6vl" podStartSLOduration=18.265764616 podStartE2EDuration="24.821029889s" podCreationTimestamp="2025-11-27 10:38:13 +0000 UTC" firstStartedPulling="2025-11-27 10:38:27.604370541 +0000 UTC m=+1012.190961305" lastFinishedPulling="2025-11-27 10:38:34.159635814 +0000 UTC m=+1018.746226578" observedRunningTime="2025-11-27 10:38:37.820460343 +0000 UTC m=+1022.407051107" watchObservedRunningTime="2025-11-27 10:38:37.821029889 +0000 UTC m=+1022.407620653" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.241002 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c9d8f7ccf-5z2jj"] Nov 27 10:38:38 crc kubenswrapper[4806]: E1127 10:38:38.241959 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b289081e-bb54-4046-9e62-37e47cacca99" containerName="placement-db-sync" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.241975 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="b289081e-bb54-4046-9e62-37e47cacca99" containerName="placement-db-sync" Nov 27 10:38:38 crc kubenswrapper[4806]: E1127 10:38:38.241984 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c109a218-41c4-4872-9fee-fbba98bb0143" containerName="keystone-bootstrap" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.241992 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c109a218-41c4-4872-9fee-fbba98bb0143" containerName="keystone-bootstrap" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.242171 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="c109a218-41c4-4872-9fee-fbba98bb0143" containerName="keystone-bootstrap" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.242187 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="b289081e-bb54-4046-9e62-37e47cacca99" containerName="placement-db-sync" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.243036 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.253848 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.254284 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.254455 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.254620 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.254748 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxgbr" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.254747 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.274137 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c9d8f7ccf-5z2jj"] Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.378339 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5689867474-qpmnv"] Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.379898 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.388915 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.389002 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.389187 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.389454 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.389484 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-sfrbs" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.392747 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5689867474-qpmnv"] Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.400737 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twvfj\" (UniqueName: \"kubernetes.io/projected/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-kube-api-access-twvfj\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.400803 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-credential-keys\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.400900 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-fernet-keys\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.400940 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-config-data\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.400995 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-scripts\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.401021 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-public-tls-certs\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.401047 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-combined-ca-bundle\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.401067 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-internal-tls-certs\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.509762 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-config-data\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.509837 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-fernet-keys\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.509867 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e42e515-17c4-4306-922d-665279ec1c3f-logs\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.509894 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-config-data\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.509939 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hffjh\" (UniqueName: \"kubernetes.io/projected/9e42e515-17c4-4306-922d-665279ec1c3f-kube-api-access-hffjh\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.509970 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-public-tls-certs\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.509988 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-scripts\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.510004 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-public-tls-certs\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.510020 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-internal-tls-certs\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.510041 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-combined-ca-bundle\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.510059 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-internal-tls-certs\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.510087 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twvfj\" (UniqueName: \"kubernetes.io/projected/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-kube-api-access-twvfj\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.510106 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-scripts\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.510126 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-credential-keys\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.510140 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-combined-ca-bundle\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.517640 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-public-tls-certs\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.528080 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-fernet-keys\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.531337 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-scripts\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.531893 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-config-data\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.533554 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-combined-ca-bundle\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.537867 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-internal-tls-certs\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.541396 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-credential-keys\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.562121 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twvfj\" (UniqueName: \"kubernetes.io/projected/e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2-kube-api-access-twvfj\") pod \"keystone-c9d8f7ccf-5z2jj\" (UID: \"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2\") " pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.611906 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-config-data\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.611977 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e42e515-17c4-4306-922d-665279ec1c3f-logs\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.612047 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hffjh\" (UniqueName: \"kubernetes.io/projected/9e42e515-17c4-4306-922d-665279ec1c3f-kube-api-access-hffjh\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.612078 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-public-tls-certs\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.612102 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-internal-tls-certs\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.612146 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-scripts\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.612174 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-combined-ca-bundle\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.617848 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e42e515-17c4-4306-922d-665279ec1c3f-logs\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.618665 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-combined-ca-bundle\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.618717 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-config-data\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.619794 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-internal-tls-certs\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.628880 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-scripts\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.629492 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.629940 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e42e515-17c4-4306-922d-665279ec1c3f-public-tls-certs\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.669526 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hffjh\" (UniqueName: \"kubernetes.io/projected/9e42e515-17c4-4306-922d-665279ec1c3f-kube-api-access-hffjh\") pod \"placement-5689867474-qpmnv\" (UID: \"9e42e515-17c4-4306-922d-665279ec1c3f\") " pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.729969 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.832169 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c76df9f57-fglss" event={"ID":"93660245-72f2-4792-af51-3ac612926729","Type":"ContainerStarted","Data":"7d1be035ab6954ac7cfbcd210e068abe6b35667c5246fd727606aa5a0625af60"} Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.832715 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c76df9f57-fglss" event={"ID":"93660245-72f2-4792-af51-3ac612926729","Type":"ContainerStarted","Data":"d669aa5700b79c103f23d73372f9640302f64fef5411712936ceee50a07764cc"} Nov 27 10:38:38 crc kubenswrapper[4806]: I1127 10:38:38.833745 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:38:39 crc kubenswrapper[4806]: I1127 10:38:39.787525 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5c76df9f57-fglss" podStartSLOduration=6.787492341 podStartE2EDuration="6.787492341s" podCreationTimestamp="2025-11-27 10:38:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:38:38.940612357 +0000 UTC m=+1023.527203121" watchObservedRunningTime="2025-11-27 10:38:39.787492341 +0000 UTC m=+1024.374083105" Nov 27 10:38:39 crc kubenswrapper[4806]: I1127 10:38:39.797386 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c9d8f7ccf-5z2jj"] Nov 27 10:38:39 crc kubenswrapper[4806]: I1127 10:38:39.884892 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cwqkj" event={"ID":"552d400b-d5b2-40c8-9654-bfc70ab5bab6","Type":"ContainerStarted","Data":"e431d6ce993c73e597a8d790fb4f7749d7911b0e9afb0c85604899710d4a58cf"} Nov 27 10:38:39 crc kubenswrapper[4806]: I1127 10:38:39.891268 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c9d8f7ccf-5z2jj" event={"ID":"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2","Type":"ContainerStarted","Data":"0956aec85ee50e493ac3c84ae3a7cee855da0d3e953984231ce962297b1fd2a8"} Nov 27 10:38:39 crc kubenswrapper[4806]: I1127 10:38:39.941751 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-cwqkj" podStartSLOduration=4.348242079 podStartE2EDuration="40.941713578s" podCreationTimestamp="2025-11-27 10:37:59 +0000 UTC" firstStartedPulling="2025-11-27 10:38:01.06056643 +0000 UTC m=+985.647157194" lastFinishedPulling="2025-11-27 10:38:37.654037929 +0000 UTC m=+1022.240628693" observedRunningTime="2025-11-27 10:38:39.935028883 +0000 UTC m=+1024.521619637" watchObservedRunningTime="2025-11-27 10:38:39.941713578 +0000 UTC m=+1024.528304342" Nov 27 10:38:40 crc kubenswrapper[4806]: I1127 10:38:40.051355 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5689867474-qpmnv"] Nov 27 10:38:40 crc kubenswrapper[4806]: W1127 10:38:40.056396 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e42e515_17c4_4306_922d_665279ec1c3f.slice/crio-46767fae46e0bdc58eeb7c5ad7488f72fc92940ec6b0c5c7cbfc81f22ebd061b WatchSource:0}: Error finding container 46767fae46e0bdc58eeb7c5ad7488f72fc92940ec6b0c5c7cbfc81f22ebd061b: Status 404 returned error can't find the container with id 46767fae46e0bdc58eeb7c5ad7488f72fc92940ec6b0c5c7cbfc81f22ebd061b Nov 27 10:38:40 crc kubenswrapper[4806]: I1127 10:38:40.904973 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c9d8f7ccf-5z2jj" event={"ID":"e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2","Type":"ContainerStarted","Data":"6257287f15e3b546c887571ef34b51defa29524b9b6a1ff87185815a96edeb95"} Nov 27 10:38:40 crc kubenswrapper[4806]: I1127 10:38:40.907106 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:38:40 crc kubenswrapper[4806]: I1127 10:38:40.915997 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5689867474-qpmnv" event={"ID":"9e42e515-17c4-4306-922d-665279ec1c3f","Type":"ContainerStarted","Data":"535b1740fa269663d9a6c6d68d1204b67fbd230a710297a934841e498adb639c"} Nov 27 10:38:40 crc kubenswrapper[4806]: I1127 10:38:40.916042 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5689867474-qpmnv" event={"ID":"9e42e515-17c4-4306-922d-665279ec1c3f","Type":"ContainerStarted","Data":"46767fae46e0bdc58eeb7c5ad7488f72fc92940ec6b0c5c7cbfc81f22ebd061b"} Nov 27 10:38:41 crc kubenswrapper[4806]: I1127 10:38:41.935936 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9vkqx" event={"ID":"11b83e17-71ca-4f23-ad17-950a5de85a12","Type":"ContainerStarted","Data":"4db8338c12c4c2627d7be068f50323ac9588babf56f5307588154ee1b2d68d02"} Nov 27 10:38:41 crc kubenswrapper[4806]: I1127 10:38:41.939163 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5689867474-qpmnv" event={"ID":"9e42e515-17c4-4306-922d-665279ec1c3f","Type":"ContainerStarted","Data":"125e71729ae7280d2a80f09534f3a4fbd46386dbe906757ac5e140896fa1efdd"} Nov 27 10:38:41 crc kubenswrapper[4806]: I1127 10:38:41.939233 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:41 crc kubenswrapper[4806]: I1127 10:38:41.939287 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5689867474-qpmnv" Nov 27 10:38:41 crc kubenswrapper[4806]: I1127 10:38:41.969799 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-c9d8f7ccf-5z2jj" podStartSLOduration=3.969773198 podStartE2EDuration="3.969773198s" podCreationTimestamp="2025-11-27 10:38:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:38:40.933112651 +0000 UTC m=+1025.519703415" watchObservedRunningTime="2025-11-27 10:38:41.969773198 +0000 UTC m=+1026.556363952" Nov 27 10:38:41 crc kubenswrapper[4806]: I1127 10:38:41.975073 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-9vkqx" podStartSLOduration=3.401931468 podStartE2EDuration="42.975045945s" podCreationTimestamp="2025-11-27 10:37:59 +0000 UTC" firstStartedPulling="2025-11-27 10:38:01.19469743 +0000 UTC m=+985.781288194" lastFinishedPulling="2025-11-27 10:38:40.767811907 +0000 UTC m=+1025.354402671" observedRunningTime="2025-11-27 10:38:41.966700143 +0000 UTC m=+1026.553290907" watchObservedRunningTime="2025-11-27 10:38:41.975045945 +0000 UTC m=+1026.561636709" Nov 27 10:38:42 crc kubenswrapper[4806]: I1127 10:38:42.026459 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5689867474-qpmnv" podStartSLOduration=4.02642058 podStartE2EDuration="4.02642058s" podCreationTimestamp="2025-11-27 10:38:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:38:42.004375239 +0000 UTC m=+1026.590966003" watchObservedRunningTime="2025-11-27 10:38:42.02642058 +0000 UTC m=+1026.613011344" Nov 27 10:38:43 crc kubenswrapper[4806]: I1127 10:38:43.650404 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:38:43 crc kubenswrapper[4806]: I1127 10:38:43.652801 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:38:44 crc kubenswrapper[4806]: I1127 10:38:44.705469 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9q6vl" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerName="registry-server" probeResult="failure" output=< Nov 27 10:38:44 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 10:38:44 crc kubenswrapper[4806]: > Nov 27 10:38:45 crc kubenswrapper[4806]: I1127 10:38:45.069404 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:38:45 crc kubenswrapper[4806]: I1127 10:38:45.156162 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw"] Nov 27 10:38:45 crc kubenswrapper[4806]: I1127 10:38:45.156443 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" podUID="68ebcee0-f362-4953-8775-1add49fb4add" containerName="dnsmasq-dns" containerID="cri-o://49bf83bfca2471a29892592b546f2f49c3dcc225bc37d76e3e0df21e7b9747d1" gracePeriod=10 Nov 27 10:38:45 crc kubenswrapper[4806]: I1127 10:38:45.719196 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" podUID="68ebcee0-f362-4953-8775-1add49fb4add" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.134:5353: connect: connection refused" Nov 27 10:38:45 crc kubenswrapper[4806]: I1127 10:38:45.980294 4806 generic.go:334] "Generic (PLEG): container finished" podID="68ebcee0-f362-4953-8775-1add49fb4add" containerID="49bf83bfca2471a29892592b546f2f49c3dcc225bc37d76e3e0df21e7b9747d1" exitCode=0 Nov 27 10:38:45 crc kubenswrapper[4806]: I1127 10:38:45.980340 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" event={"ID":"68ebcee0-f362-4953-8775-1add49fb4add","Type":"ContainerDied","Data":"49bf83bfca2471a29892592b546f2f49c3dcc225bc37d76e3e0df21e7b9747d1"} Nov 27 10:38:49 crc kubenswrapper[4806]: I1127 10:38:49.019972 4806 generic.go:334] "Generic (PLEG): container finished" podID="11b83e17-71ca-4f23-ad17-950a5de85a12" containerID="4db8338c12c4c2627d7be068f50323ac9588babf56f5307588154ee1b2d68d02" exitCode=0 Nov 27 10:38:49 crc kubenswrapper[4806]: I1127 10:38:49.020041 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9vkqx" event={"ID":"11b83e17-71ca-4f23-ad17-950a5de85a12","Type":"ContainerDied","Data":"4db8338c12c4c2627d7be068f50323ac9588babf56f5307588154ee1b2d68d02"} Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:50.720043 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" podUID="68ebcee0-f362-4953-8775-1add49fb4add" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.134:5353: connect: connection refused" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.103433 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9vkqx" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.110606 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9vkqx" event={"ID":"11b83e17-71ca-4f23-ad17-950a5de85a12","Type":"ContainerDied","Data":"56f7d94746a17b9cd17a006fddd49da69b3e6c38fde001caea0baf9ab370c0f5"} Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.110652 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56f7d94746a17b9cd17a006fddd49da69b3e6c38fde001caea0baf9ab370c0f5" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.110725 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9vkqx" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.118844 4806 generic.go:334] "Generic (PLEG): container finished" podID="552d400b-d5b2-40c8-9654-bfc70ab5bab6" containerID="e431d6ce993c73e597a8d790fb4f7749d7911b0e9afb0c85604899710d4a58cf" exitCode=0 Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.120267 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cwqkj" event={"ID":"552d400b-d5b2-40c8-9654-bfc70ab5bab6","Type":"ContainerDied","Data":"e431d6ce993c73e597a8d790fb4f7749d7911b0e9afb0c85604899710d4a58cf"} Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.139601 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzbjf\" (UniqueName: \"kubernetes.io/projected/11b83e17-71ca-4f23-ad17-950a5de85a12-kube-api-access-pzbjf\") pod \"11b83e17-71ca-4f23-ad17-950a5de85a12\" (UID: \"11b83e17-71ca-4f23-ad17-950a5de85a12\") " Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.139913 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11b83e17-71ca-4f23-ad17-950a5de85a12-db-sync-config-data\") pod \"11b83e17-71ca-4f23-ad17-950a5de85a12\" (UID: \"11b83e17-71ca-4f23-ad17-950a5de85a12\") " Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.140058 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b83e17-71ca-4f23-ad17-950a5de85a12-combined-ca-bundle\") pod \"11b83e17-71ca-4f23-ad17-950a5de85a12\" (UID: \"11b83e17-71ca-4f23-ad17-950a5de85a12\") " Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.154564 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11b83e17-71ca-4f23-ad17-950a5de85a12-kube-api-access-pzbjf" (OuterVolumeSpecName: "kube-api-access-pzbjf") pod "11b83e17-71ca-4f23-ad17-950a5de85a12" (UID: "11b83e17-71ca-4f23-ad17-950a5de85a12"). InnerVolumeSpecName "kube-api-access-pzbjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.207652 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b83e17-71ca-4f23-ad17-950a5de85a12-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "11b83e17-71ca-4f23-ad17-950a5de85a12" (UID: "11b83e17-71ca-4f23-ad17-950a5de85a12"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.237767 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11b83e17-71ca-4f23-ad17-950a5de85a12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11b83e17-71ca-4f23-ad17-950a5de85a12" (UID: "11b83e17-71ca-4f23-ad17-950a5de85a12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.247758 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzbjf\" (UniqueName: \"kubernetes.io/projected/11b83e17-71ca-4f23-ad17-950a5de85a12-kube-api-access-pzbjf\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.247790 4806 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11b83e17-71ca-4f23-ad17-950a5de85a12-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.247802 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b83e17-71ca-4f23-ad17-950a5de85a12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.427450 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.556341 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-ovsdbserver-nb\") pod \"68ebcee0-f362-4953-8775-1add49fb4add\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.556643 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9rln\" (UniqueName: \"kubernetes.io/projected/68ebcee0-f362-4953-8775-1add49fb4add-kube-api-access-p9rln\") pod \"68ebcee0-f362-4953-8775-1add49fb4add\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.556780 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-dns-svc\") pod \"68ebcee0-f362-4953-8775-1add49fb4add\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.556843 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-config\") pod \"68ebcee0-f362-4953-8775-1add49fb4add\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.556906 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-ovsdbserver-sb\") pod \"68ebcee0-f362-4953-8775-1add49fb4add\" (UID: \"68ebcee0-f362-4953-8775-1add49fb4add\") " Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.561476 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68ebcee0-f362-4953-8775-1add49fb4add-kube-api-access-p9rln" (OuterVolumeSpecName: "kube-api-access-p9rln") pod "68ebcee0-f362-4953-8775-1add49fb4add" (UID: "68ebcee0-f362-4953-8775-1add49fb4add"). InnerVolumeSpecName "kube-api-access-p9rln". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.603177 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "68ebcee0-f362-4953-8775-1add49fb4add" (UID: "68ebcee0-f362-4953-8775-1add49fb4add"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.606263 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "68ebcee0-f362-4953-8775-1add49fb4add" (UID: "68ebcee0-f362-4953-8775-1add49fb4add"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.616754 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-config" (OuterVolumeSpecName: "config") pod "68ebcee0-f362-4953-8775-1add49fb4add" (UID: "68ebcee0-f362-4953-8775-1add49fb4add"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.629600 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "68ebcee0-f362-4953-8775-1add49fb4add" (UID: "68ebcee0-f362-4953-8775-1add49fb4add"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.659476 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.659509 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9rln\" (UniqueName: \"kubernetes.io/projected/68ebcee0-f362-4953-8775-1add49fb4add-kube-api-access-p9rln\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.659528 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.659537 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:51 crc kubenswrapper[4806]: I1127 10:38:51.659545 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68ebcee0-f362-4953-8775-1add49fb4add-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.132201 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05ff370-09c4-45c1-b101-a6b277a34d3d","Type":"ContainerStarted","Data":"73aeda0d2b661d85a701e6d5162caafce27cfad08c384b94650d9757e3f05fee"} Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.132888 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.132648 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="proxy-httpd" containerID="cri-o://73aeda0d2b661d85a701e6d5162caafce27cfad08c384b94650d9757e3f05fee" gracePeriod=30 Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.132340 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="ceilometer-central-agent" containerID="cri-o://07377df1241e43c09f0052ed513662c35a9c3b54c8608fdc9491c4e98408cdf1" gracePeriod=30 Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.132667 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="sg-core" containerID="cri-o://183d9372b53f6427808ac6d0ce282b15607ed7035e79538be935f9fddf1eb346" gracePeriod=30 Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.132697 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="ceilometer-notification-agent" containerID="cri-o://9c03a47efe11a0973c4aaa2985a4039e32eb53c1fe4d3d8393c24291624b7bfb" gracePeriod=30 Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.136747 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" event={"ID":"68ebcee0-f362-4953-8775-1add49fb4add","Type":"ContainerDied","Data":"82d11378477f383061a1a868603a3737c42b192c785ff61fc8cba2df1a656ef7"} Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.136820 4806 scope.go:117] "RemoveContainer" containerID="49bf83bfca2471a29892592b546f2f49c3dcc225bc37d76e3e0df21e7b9747d1" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.136823 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.206174 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.677082148 podStartE2EDuration="53.206156525s" podCreationTimestamp="2025-11-27 10:37:59 +0000 UTC" firstStartedPulling="2025-11-27 10:38:01.696409213 +0000 UTC m=+986.282999977" lastFinishedPulling="2025-11-27 10:38:51.22548359 +0000 UTC m=+1035.812074354" observedRunningTime="2025-11-27 10:38:52.189679098 +0000 UTC m=+1036.776269872" watchObservedRunningTime="2025-11-27 10:38:52.206156525 +0000 UTC m=+1036.792747289" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.208041 4806 scope.go:117] "RemoveContainer" containerID="c0629d030f4cbf04c370623ea92033a5f6772e789f55577b4c894bcbe3680479" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.208907 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw"] Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.230738 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-8l2bw"] Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.427331 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6fb499c486-z7b8m"] Nov 27 10:38:52 crc kubenswrapper[4806]: E1127 10:38:52.427975 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ebcee0-f362-4953-8775-1add49fb4add" containerName="init" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.427990 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ebcee0-f362-4953-8775-1add49fb4add" containerName="init" Nov 27 10:38:52 crc kubenswrapper[4806]: E1127 10:38:52.428005 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ebcee0-f362-4953-8775-1add49fb4add" containerName="dnsmasq-dns" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.428012 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ebcee0-f362-4953-8775-1add49fb4add" containerName="dnsmasq-dns" Nov 27 10:38:52 crc kubenswrapper[4806]: E1127 10:38:52.428047 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b83e17-71ca-4f23-ad17-950a5de85a12" containerName="barbican-db-sync" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.428056 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b83e17-71ca-4f23-ad17-950a5de85a12" containerName="barbican-db-sync" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.428289 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="68ebcee0-f362-4953-8775-1add49fb4add" containerName="dnsmasq-dns" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.428311 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="11b83e17-71ca-4f23-ad17-950a5de85a12" containerName="barbican-db-sync" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.429644 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.436888 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hdrsj" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.437162 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.440953 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.442124 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6fb499c486-z7b8m"] Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.462308 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5bb7f69687-8mg9f"] Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.463984 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.467469 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.476786 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5bb7f69687-8mg9f"] Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.479102 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad03fa3a-e8e2-480d-8980-8a54cd70debb-config-data\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.479216 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad03fa3a-e8e2-480d-8980-8a54cd70debb-combined-ca-bundle\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.479254 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad03fa3a-e8e2-480d-8980-8a54cd70debb-logs\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.479283 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq5hs\" (UniqueName: \"kubernetes.io/projected/ad03fa3a-e8e2-480d-8980-8a54cd70debb-kube-api-access-bq5hs\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.479326 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad03fa3a-e8e2-480d-8980-8a54cd70debb-config-data-custom\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.589576 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a5920ca-ee03-41f8-9886-2c609254e7f7-config-data\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.589987 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad03fa3a-e8e2-480d-8980-8a54cd70debb-config-data-custom\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.590222 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a5920ca-ee03-41f8-9886-2c609254e7f7-logs\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.590273 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad03fa3a-e8e2-480d-8980-8a54cd70debb-config-data\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.590313 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a5920ca-ee03-41f8-9886-2c609254e7f7-combined-ca-bundle\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.590426 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad03fa3a-e8e2-480d-8980-8a54cd70debb-combined-ca-bundle\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.590450 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad03fa3a-e8e2-480d-8980-8a54cd70debb-logs\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.590479 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq5hs\" (UniqueName: \"kubernetes.io/projected/ad03fa3a-e8e2-480d-8980-8a54cd70debb-kube-api-access-bq5hs\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.590501 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9t7t\" (UniqueName: \"kubernetes.io/projected/2a5920ca-ee03-41f8-9886-2c609254e7f7-kube-api-access-n9t7t\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.590519 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a5920ca-ee03-41f8-9886-2c609254e7f7-config-data-custom\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.591548 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad03fa3a-e8e2-480d-8980-8a54cd70debb-logs\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.598903 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad03fa3a-e8e2-480d-8980-8a54cd70debb-combined-ca-bundle\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.604713 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad03fa3a-e8e2-480d-8980-8a54cd70debb-config-data-custom\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.613698 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad03fa3a-e8e2-480d-8980-8a54cd70debb-config-data\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.635915 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq5hs\" (UniqueName: \"kubernetes.io/projected/ad03fa3a-e8e2-480d-8980-8a54cd70debb-kube-api-access-bq5hs\") pod \"barbican-keystone-listener-6fb499c486-z7b8m\" (UID: \"ad03fa3a-e8e2-480d-8980-8a54cd70debb\") " pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.639901 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-869f779d85-qsl7g"] Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.641453 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.648313 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-qsl7g"] Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.688632 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.693811 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a5920ca-ee03-41f8-9886-2c609254e7f7-config-data\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.693960 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.694004 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.694064 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a5920ca-ee03-41f8-9886-2c609254e7f7-logs\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.694133 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a5920ca-ee03-41f8-9886-2c609254e7f7-combined-ca-bundle\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.694220 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x4hv\" (UniqueName: \"kubernetes.io/projected/91b270f0-0dec-4c48-ac15-dc748b910066-kube-api-access-5x4hv\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.694283 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-dns-svc\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.694353 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9t7t\" (UniqueName: \"kubernetes.io/projected/2a5920ca-ee03-41f8-9886-2c609254e7f7-kube-api-access-n9t7t\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.694387 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a5920ca-ee03-41f8-9886-2c609254e7f7-config-data-custom\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.694438 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-config\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.695617 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a5920ca-ee03-41f8-9886-2c609254e7f7-logs\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.699289 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a5920ca-ee03-41f8-9886-2c609254e7f7-combined-ca-bundle\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.712956 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a5920ca-ee03-41f8-9886-2c609254e7f7-config-data-custom\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.714582 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a5920ca-ee03-41f8-9886-2c609254e7f7-config-data\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.738431 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9t7t\" (UniqueName: \"kubernetes.io/projected/2a5920ca-ee03-41f8-9886-2c609254e7f7-kube-api-access-n9t7t\") pod \"barbican-worker-5bb7f69687-8mg9f\" (UID: \"2a5920ca-ee03-41f8-9886-2c609254e7f7\") " pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.797173 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-scripts\") pod \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.797283 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/552d400b-d5b2-40c8-9654-bfc70ab5bab6-etc-machine-id\") pod \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.797398 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-db-sync-config-data\") pod \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.797430 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-config-data\") pod \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.797448 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-combined-ca-bundle\") pod \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.797550 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnt7d\" (UniqueName: \"kubernetes.io/projected/552d400b-d5b2-40c8-9654-bfc70ab5bab6-kube-api-access-lnt7d\") pod \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\" (UID: \"552d400b-d5b2-40c8-9654-bfc70ab5bab6\") " Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.797803 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x4hv\" (UniqueName: \"kubernetes.io/projected/91b270f0-0dec-4c48-ac15-dc748b910066-kube-api-access-5x4hv\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.797829 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-dns-svc\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.797868 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-config\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.797934 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.797965 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.798878 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.800769 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-config\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.801855 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/552d400b-d5b2-40c8-9654-bfc70ab5bab6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "552d400b-d5b2-40c8-9654-bfc70ab5bab6" (UID: "552d400b-d5b2-40c8-9654-bfc70ab5bab6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.802058 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.808325 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-dns-svc\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.808984 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/552d400b-d5b2-40c8-9654-bfc70ab5bab6-kube-api-access-lnt7d" (OuterVolumeSpecName: "kube-api-access-lnt7d") pod "552d400b-d5b2-40c8-9654-bfc70ab5bab6" (UID: "552d400b-d5b2-40c8-9654-bfc70ab5bab6"). InnerVolumeSpecName "kube-api-access-lnt7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.812728 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-scripts" (OuterVolumeSpecName: "scripts") pod "552d400b-d5b2-40c8-9654-bfc70ab5bab6" (UID: "552d400b-d5b2-40c8-9654-bfc70ab5bab6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.812788 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "552d400b-d5b2-40c8-9654-bfc70ab5bab6" (UID: "552d400b-d5b2-40c8-9654-bfc70ab5bab6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.820395 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-8ccc54b-kzmfd"] Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.820630 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" Nov 27 10:38:52 crc kubenswrapper[4806]: E1127 10:38:52.820818 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="552d400b-d5b2-40c8-9654-bfc70ab5bab6" containerName="cinder-db-sync" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.820843 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="552d400b-d5b2-40c8-9654-bfc70ab5bab6" containerName="cinder-db-sync" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.821073 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="552d400b-d5b2-40c8-9654-bfc70ab5bab6" containerName="cinder-db-sync" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.822057 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.832800 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.859134 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5bb7f69687-8mg9f" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.882696 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8ccc54b-kzmfd"] Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.894661 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "552d400b-d5b2-40c8-9654-bfc70ab5bab6" (UID: "552d400b-d5b2-40c8-9654-bfc70ab5bab6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.907148 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbm4t\" (UniqueName: \"kubernetes.io/projected/3967a997-3385-4163-97e8-fe3ef5eaf57c-kube-api-access-xbm4t\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.907203 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3967a997-3385-4163-97e8-fe3ef5eaf57c-logs\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.907319 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-config-data-custom\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.907381 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-config-data\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.907401 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-combined-ca-bundle\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.907472 4806 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.907485 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.907498 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnt7d\" (UniqueName: \"kubernetes.io/projected/552d400b-d5b2-40c8-9654-bfc70ab5bab6-kube-api-access-lnt7d\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.907512 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.907523 4806 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/552d400b-d5b2-40c8-9654-bfc70ab5bab6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.926930 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x4hv\" (UniqueName: \"kubernetes.io/projected/91b270f0-0dec-4c48-ac15-dc748b910066-kube-api-access-5x4hv\") pod \"dnsmasq-dns-869f779d85-qsl7g\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:52 crc kubenswrapper[4806]: I1127 10:38:52.981207 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.010852 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-config-data-custom\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.011004 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-config-data\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.011030 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-combined-ca-bundle\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.011101 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbm4t\" (UniqueName: \"kubernetes.io/projected/3967a997-3385-4163-97e8-fe3ef5eaf57c-kube-api-access-xbm4t\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.011146 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3967a997-3385-4163-97e8-fe3ef5eaf57c-logs\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.013683 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3967a997-3385-4163-97e8-fe3ef5eaf57c-logs\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.018915 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-config-data-custom\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.022474 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-config-data\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.022962 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-combined-ca-bundle\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.035418 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-config-data" (OuterVolumeSpecName: "config-data") pod "552d400b-d5b2-40c8-9654-bfc70ab5bab6" (UID: "552d400b-d5b2-40c8-9654-bfc70ab5bab6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.041462 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbm4t\" (UniqueName: \"kubernetes.io/projected/3967a997-3385-4163-97e8-fe3ef5eaf57c-kube-api-access-xbm4t\") pod \"barbican-api-8ccc54b-kzmfd\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.116438 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/552d400b-d5b2-40c8-9654-bfc70ab5bab6-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.183021 4806 generic.go:334] "Generic (PLEG): container finished" podID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerID="73aeda0d2b661d85a701e6d5162caafce27cfad08c384b94650d9757e3f05fee" exitCode=0 Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.183054 4806 generic.go:334] "Generic (PLEG): container finished" podID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerID="183d9372b53f6427808ac6d0ce282b15607ed7035e79538be935f9fddf1eb346" exitCode=2 Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.183097 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05ff370-09c4-45c1-b101-a6b277a34d3d","Type":"ContainerDied","Data":"73aeda0d2b661d85a701e6d5162caafce27cfad08c384b94650d9757e3f05fee"} Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.183124 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05ff370-09c4-45c1-b101-a6b277a34d3d","Type":"ContainerDied","Data":"183d9372b53f6427808ac6d0ce282b15607ed7035e79538be935f9fddf1eb346"} Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.212208 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cwqkj" event={"ID":"552d400b-d5b2-40c8-9654-bfc70ab5bab6","Type":"ContainerDied","Data":"7b53b5c38a61c2b7430cfbff74b42faaeec94ad45cf3647384d4c18dfd638a69"} Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.212267 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b53b5c38a61c2b7430cfbff74b42faaeec94ad45cf3647384d4c18dfd638a69" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.212649 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cwqkj" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.446859 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:53 crc kubenswrapper[4806]: E1127 10:38:53.687644 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd05ff370_09c4_45c1_b101_a6b277a34d3d.slice/crio-07377df1241e43c09f0052ed513662c35a9c3b54c8608fdc9491c4e98408cdf1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd05ff370_09c4_45c1_b101_a6b277a34d3d.slice/crio-conmon-07377df1241e43c09f0052ed513662c35a9c3b54c8608fdc9491c4e98408cdf1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod552d400b_d5b2_40c8_9654_bfc70ab5bab6.slice/crio-7b53b5c38a61c2b7430cfbff74b42faaeec94ad45cf3647384d4c18dfd638a69\": RecentStats: unable to find data in memory cache]" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.755526 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.776336 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.789473 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.793417 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.793687 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5jcsw" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.793972 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.794231 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.838169 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.838597 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-scripts\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.838663 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-config-data\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.838707 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb4xs\" (UniqueName: \"kubernetes.io/projected/070121ad-363b-4c59-b6ac-47776d46fb25-kube-api-access-rb4xs\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.838743 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/070121ad-363b-4c59-b6ac-47776d46fb25-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.838777 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.938481 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-qsl7g"] Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.940494 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-scripts\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.940569 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-config-data\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.940615 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb4xs\" (UniqueName: \"kubernetes.io/projected/070121ad-363b-4c59-b6ac-47776d46fb25-kube-api-access-rb4xs\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.940659 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/070121ad-363b-4c59-b6ac-47776d46fb25-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.940687 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.940757 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.948053 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/070121ad-363b-4c59-b6ac-47776d46fb25-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.952338 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-scripts\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.960145 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.969891 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:53 crc kubenswrapper[4806]: I1127 10:38:53.978022 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-config-data\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.001098 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb4xs\" (UniqueName: \"kubernetes.io/projected/070121ad-363b-4c59-b6ac-47776d46fb25-kube-api-access-rb4xs\") pod \"cinder-scheduler-0\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " pod="openstack/cinder-scheduler-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.095948 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6fb499c486-z7b8m"] Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.152950 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68ebcee0-f362-4953-8775-1add49fb4add" path="/var/lib/kubelet/pods/68ebcee0-f362-4953-8775-1add49fb4add/volumes" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.153764 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-48ht6"] Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.155749 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.173162 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.198739 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-48ht6"] Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.275796 4806 generic.go:334] "Generic (PLEG): container finished" podID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerID="07377df1241e43c09f0052ed513662c35a9c3b54c8608fdc9491c4e98408cdf1" exitCode=0 Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.275889 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05ff370-09c4-45c1-b101-a6b277a34d3d","Type":"ContainerDied","Data":"07377df1241e43c09f0052ed513662c35a9c3b54c8608fdc9491c4e98408cdf1"} Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.280284 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" event={"ID":"ad03fa3a-e8e2-480d-8980-8a54cd70debb","Type":"ContainerStarted","Data":"c4e29f556d2cba38987f3ec25bb18a905396cb475ba8c45bf3e22661f8210fc1"} Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.310023 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5bb7f69687-8mg9f"] Nov 27 10:38:54 crc kubenswrapper[4806]: W1127 10:38:54.330605 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a5920ca_ee03_41f8_9886_2c609254e7f7.slice/crio-926771c6c43a20dff6ca374d9d0d413c9e190bbca5fe6b87348258dfd650fb00 WatchSource:0}: Error finding container 926771c6c43a20dff6ca374d9d0d413c9e190bbca5fe6b87348258dfd650fb00: Status 404 returned error can't find the container with id 926771c6c43a20dff6ca374d9d0d413c9e190bbca5fe6b87348258dfd650fb00 Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.354821 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.355024 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-config\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.355096 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-dns-svc\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.355490 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgccm\" (UniqueName: \"kubernetes.io/projected/4f7a5d92-530d-4675-9a1b-6cf277064169-kube-api-access-lgccm\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.355564 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.413473 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.414984 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.421637 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.423941 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.460758 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.460924 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.461048 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-config\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.461082 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-dns-svc\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.461162 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgccm\" (UniqueName: \"kubernetes.io/projected/4f7a5d92-530d-4675-9a1b-6cf277064169-kube-api-access-lgccm\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.462677 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.463308 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-dns-svc\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.464653 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.473451 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-config\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.505946 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgccm\" (UniqueName: \"kubernetes.io/projected/4f7a5d92-530d-4675-9a1b-6cf277064169-kube-api-access-lgccm\") pod \"dnsmasq-dns-58db5546cc-48ht6\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.564195 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f3feff2-b2d0-42b7-9adc-e65686db77f1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.564317 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f3feff2-b2d0-42b7-9adc-e65686db77f1-logs\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.564352 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-config-data\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.564397 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.564423 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-config-data-custom\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.564458 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-scripts\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.564500 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2fcf\" (UniqueName: \"kubernetes.io/projected/3f3feff2-b2d0-42b7-9adc-e65686db77f1-kube-api-access-g2fcf\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.669315 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.669672 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-config-data-custom\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.669717 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-scripts\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.669757 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2fcf\" (UniqueName: \"kubernetes.io/projected/3f3feff2-b2d0-42b7-9adc-e65686db77f1-kube-api-access-g2fcf\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.669782 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f3feff2-b2d0-42b7-9adc-e65686db77f1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.669837 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f3feff2-b2d0-42b7-9adc-e65686db77f1-logs\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.669868 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-config-data\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.671689 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f3feff2-b2d0-42b7-9adc-e65686db77f1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.672825 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-qsl7g"] Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.673792 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f3feff2-b2d0-42b7-9adc-e65686db77f1-logs\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.678441 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.678473 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-scripts\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.683447 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-config-data\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.687977 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-config-data-custom\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.688646 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.693945 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2fcf\" (UniqueName: \"kubernetes.io/projected/3f3feff2-b2d0-42b7-9adc-e65686db77f1-kube-api-access-g2fcf\") pod \"cinder-api-0\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.773042 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.776695 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9q6vl" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerName="registry-server" probeResult="failure" output=< Nov 27 10:38:54 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 10:38:54 crc kubenswrapper[4806]: > Nov 27 10:38:54 crc kubenswrapper[4806]: I1127 10:38:54.814722 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8ccc54b-kzmfd"] Nov 27 10:38:54 crc kubenswrapper[4806]: W1127 10:38:54.847898 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3967a997_3385_4163_97e8_fe3ef5eaf57c.slice/crio-c4ca8251cf5920a77392245356caa8224b5dfd9d0ca3fd3da897f2961c6334d3 WatchSource:0}: Error finding container c4ca8251cf5920a77392245356caa8224b5dfd9d0ca3fd3da897f2961c6334d3: Status 404 returned error can't find the container with id c4ca8251cf5920a77392245356caa8224b5dfd9d0ca3fd3da897f2961c6334d3 Nov 27 10:38:55 crc kubenswrapper[4806]: I1127 10:38:55.038449 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 10:38:55 crc kubenswrapper[4806]: W1127 10:38:55.042135 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod070121ad_363b_4c59_b6ac_47776d46fb25.slice/crio-001528992b8a04707b3c48dc9f0efb0de30217fe0bfe41f8d302412e1ed8befc WatchSource:0}: Error finding container 001528992b8a04707b3c48dc9f0efb0de30217fe0bfe41f8d302412e1ed8befc: Status 404 returned error can't find the container with id 001528992b8a04707b3c48dc9f0efb0de30217fe0bfe41f8d302412e1ed8befc Nov 27 10:38:55 crc kubenswrapper[4806]: I1127 10:38:55.328217 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8ccc54b-kzmfd" event={"ID":"3967a997-3385-4163-97e8-fe3ef5eaf57c","Type":"ContainerStarted","Data":"6e514a446147a41fe562cb099f03806ccf4f5dbd165ea55189506f356e536b89"} Nov 27 10:38:55 crc kubenswrapper[4806]: I1127 10:38:55.328913 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8ccc54b-kzmfd" event={"ID":"3967a997-3385-4163-97e8-fe3ef5eaf57c","Type":"ContainerStarted","Data":"c4ca8251cf5920a77392245356caa8224b5dfd9d0ca3fd3da897f2961c6334d3"} Nov 27 10:38:55 crc kubenswrapper[4806]: I1127 10:38:55.332683 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-qsl7g" event={"ID":"91b270f0-0dec-4c48-ac15-dc748b910066","Type":"ContainerStarted","Data":"5738af748aa7994b002c9ae1cfda1a4bdf3d975ac5fecff838162e4cfc1f2e92"} Nov 27 10:38:55 crc kubenswrapper[4806]: I1127 10:38:55.333021 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-qsl7g" event={"ID":"91b270f0-0dec-4c48-ac15-dc748b910066","Type":"ContainerStarted","Data":"6bac495a9eb95e9f41ae620e79fc279ad6e2588e2dc1469be17acd1fe93b539c"} Nov 27 10:38:55 crc kubenswrapper[4806]: I1127 10:38:55.346055 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5bb7f69687-8mg9f" event={"ID":"2a5920ca-ee03-41f8-9886-2c609254e7f7","Type":"ContainerStarted","Data":"926771c6c43a20dff6ca374d9d0d413c9e190bbca5fe6b87348258dfd650fb00"} Nov 27 10:38:55 crc kubenswrapper[4806]: I1127 10:38:55.349877 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"070121ad-363b-4c59-b6ac-47776d46fb25","Type":"ContainerStarted","Data":"001528992b8a04707b3c48dc9f0efb0de30217fe0bfe41f8d302412e1ed8befc"} Nov 27 10:38:55 crc kubenswrapper[4806]: I1127 10:38:55.452989 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 27 10:38:55 crc kubenswrapper[4806]: I1127 10:38:55.606295 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-48ht6"] Nov 27 10:38:55 crc kubenswrapper[4806]: W1127 10:38:55.668323 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f7a5d92_530d_4675_9a1b_6cf277064169.slice/crio-4ef388f41813a863e716071dae4548acf05379b26d07a2fa2583fcbac1c7cf00 WatchSource:0}: Error finding container 4ef388f41813a863e716071dae4548acf05379b26d07a2fa2583fcbac1c7cf00: Status 404 returned error can't find the container with id 4ef388f41813a863e716071dae4548acf05379b26d07a2fa2583fcbac1c7cf00 Nov 27 10:38:55 crc kubenswrapper[4806]: I1127 10:38:55.823416 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:55 crc kubenswrapper[4806]: I1127 10:38:55.999858 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-ovsdbserver-nb\") pod \"91b270f0-0dec-4c48-ac15-dc748b910066\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " Nov 27 10:38:55 crc kubenswrapper[4806]: I1127 10:38:55.999924 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-ovsdbserver-sb\") pod \"91b270f0-0dec-4c48-ac15-dc748b910066\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " Nov 27 10:38:55 crc kubenswrapper[4806]: I1127 10:38:55.999975 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x4hv\" (UniqueName: \"kubernetes.io/projected/91b270f0-0dec-4c48-ac15-dc748b910066-kube-api-access-5x4hv\") pod \"91b270f0-0dec-4c48-ac15-dc748b910066\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.000021 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-config\") pod \"91b270f0-0dec-4c48-ac15-dc748b910066\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.000050 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-dns-svc\") pod \"91b270f0-0dec-4c48-ac15-dc748b910066\" (UID: \"91b270f0-0dec-4c48-ac15-dc748b910066\") " Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.006311 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91b270f0-0dec-4c48-ac15-dc748b910066-kube-api-access-5x4hv" (OuterVolumeSpecName: "kube-api-access-5x4hv") pod "91b270f0-0dec-4c48-ac15-dc748b910066" (UID: "91b270f0-0dec-4c48-ac15-dc748b910066"). InnerVolumeSpecName "kube-api-access-5x4hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.026458 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "91b270f0-0dec-4c48-ac15-dc748b910066" (UID: "91b270f0-0dec-4c48-ac15-dc748b910066"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.027870 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "91b270f0-0dec-4c48-ac15-dc748b910066" (UID: "91b270f0-0dec-4c48-ac15-dc748b910066"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.036345 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-config" (OuterVolumeSpecName: "config") pod "91b270f0-0dec-4c48-ac15-dc748b910066" (UID: "91b270f0-0dec-4c48-ac15-dc748b910066"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.043520 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "91b270f0-0dec-4c48-ac15-dc748b910066" (UID: "91b270f0-0dec-4c48-ac15-dc748b910066"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.101715 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.102137 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.102152 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x4hv\" (UniqueName: \"kubernetes.io/projected/91b270f0-0dec-4c48-ac15-dc748b910066-kube-api-access-5x4hv\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.102164 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.102177 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91b270f0-0dec-4c48-ac15-dc748b910066-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.411180 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3f3feff2-b2d0-42b7-9adc-e65686db77f1","Type":"ContainerStarted","Data":"135864be863d704c327f5aec71c6b78303cfe46461261a72e5e41562d08b6122"} Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.444806 4806 generic.go:334] "Generic (PLEG): container finished" podID="4f7a5d92-530d-4675-9a1b-6cf277064169" containerID="41e57d921ac6d6ea0b57795152d1e3c5dcc72a7ad4044801b861fc943187a730" exitCode=0 Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.444880 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-48ht6" event={"ID":"4f7a5d92-530d-4675-9a1b-6cf277064169","Type":"ContainerDied","Data":"41e57d921ac6d6ea0b57795152d1e3c5dcc72a7ad4044801b861fc943187a730"} Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.444907 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-48ht6" event={"ID":"4f7a5d92-530d-4675-9a1b-6cf277064169","Type":"ContainerStarted","Data":"4ef388f41813a863e716071dae4548acf05379b26d07a2fa2583fcbac1c7cf00"} Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.451891 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8ccc54b-kzmfd" event={"ID":"3967a997-3385-4163-97e8-fe3ef5eaf57c","Type":"ContainerStarted","Data":"e933eeb67d80d9dbf4b5a96b9e24bfb6ca8410b0b02cb6ba88b3f0b57ad11d52"} Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.451947 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.451976 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.457945 4806 generic.go:334] "Generic (PLEG): container finished" podID="91b270f0-0dec-4c48-ac15-dc748b910066" containerID="5738af748aa7994b002c9ae1cfda1a4bdf3d975ac5fecff838162e4cfc1f2e92" exitCode=0 Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.458022 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-qsl7g" event={"ID":"91b270f0-0dec-4c48-ac15-dc748b910066","Type":"ContainerDied","Data":"5738af748aa7994b002c9ae1cfda1a4bdf3d975ac5fecff838162e4cfc1f2e92"} Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.458066 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-qsl7g" event={"ID":"91b270f0-0dec-4c48-ac15-dc748b910066","Type":"ContainerDied","Data":"6bac495a9eb95e9f41ae620e79fc279ad6e2588e2dc1469be17acd1fe93b539c"} Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.458088 4806 scope.go:117] "RemoveContainer" containerID="5738af748aa7994b002c9ae1cfda1a4bdf3d975ac5fecff838162e4cfc1f2e92" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.458379 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-qsl7g" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.498847 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-8ccc54b-kzmfd" podStartSLOduration=4.4988286760000005 podStartE2EDuration="4.498828676s" podCreationTimestamp="2025-11-27 10:38:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:38:56.496635296 +0000 UTC m=+1041.083226070" watchObservedRunningTime="2025-11-27 10:38:56.498828676 +0000 UTC m=+1041.085419440" Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.555304 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-qsl7g"] Nov 27 10:38:56 crc kubenswrapper[4806]: I1127 10:38:56.558421 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-qsl7g"] Nov 27 10:38:57 crc kubenswrapper[4806]: I1127 10:38:57.419309 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 27 10:38:57 crc kubenswrapper[4806]: I1127 10:38:57.478900 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3f3feff2-b2d0-42b7-9adc-e65686db77f1","Type":"ContainerStarted","Data":"34c085f1489f8bbc3e84ba7457adeea813450610fa09f58cac40df9a2b288c95"} Nov 27 10:38:57 crc kubenswrapper[4806]: I1127 10:38:57.505555 4806 generic.go:334] "Generic (PLEG): container finished" podID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerID="9c03a47efe11a0973c4aaa2985a4039e32eb53c1fe4d3d8393c24291624b7bfb" exitCode=0 Nov 27 10:38:57 crc kubenswrapper[4806]: I1127 10:38:57.506884 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05ff370-09c4-45c1-b101-a6b277a34d3d","Type":"ContainerDied","Data":"9c03a47efe11a0973c4aaa2985a4039e32eb53c1fe4d3d8393c24291624b7bfb"} Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.029347 4806 scope.go:117] "RemoveContainer" containerID="5738af748aa7994b002c9ae1cfda1a4bdf3d975ac5fecff838162e4cfc1f2e92" Nov 27 10:38:58 crc kubenswrapper[4806]: E1127 10:38:58.031087 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5738af748aa7994b002c9ae1cfda1a4bdf3d975ac5fecff838162e4cfc1f2e92\": container with ID starting with 5738af748aa7994b002c9ae1cfda1a4bdf3d975ac5fecff838162e4cfc1f2e92 not found: ID does not exist" containerID="5738af748aa7994b002c9ae1cfda1a4bdf3d975ac5fecff838162e4cfc1f2e92" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.031158 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5738af748aa7994b002c9ae1cfda1a4bdf3d975ac5fecff838162e4cfc1f2e92"} err="failed to get container status \"5738af748aa7994b002c9ae1cfda1a4bdf3d975ac5fecff838162e4cfc1f2e92\": rpc error: code = NotFound desc = could not find container \"5738af748aa7994b002c9ae1cfda1a4bdf3d975ac5fecff838162e4cfc1f2e92\": container with ID starting with 5738af748aa7994b002c9ae1cfda1a4bdf3d975ac5fecff838162e4cfc1f2e92 not found: ID does not exist" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.186279 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91b270f0-0dec-4c48-ac15-dc748b910066" path="/var/lib/kubelet/pods/91b270f0-0dec-4c48-ac15-dc748b910066/volumes" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.527501 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-48ht6" event={"ID":"4f7a5d92-530d-4675-9a1b-6cf277064169","Type":"ContainerStarted","Data":"b9516b0292adbf0cd5fdac0d4a4170075d42dd3fd6674be640f04191a1621962"} Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.529691 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.591041 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58db5546cc-48ht6" podStartSLOduration=4.591011191 podStartE2EDuration="4.591011191s" podCreationTimestamp="2025-11-27 10:38:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:38:58.568607574 +0000 UTC m=+1043.155198338" watchObservedRunningTime="2025-11-27 10:38:58.591011191 +0000 UTC m=+1043.177601955" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.699698 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.786377 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05ff370-09c4-45c1-b101-a6b277a34d3d-run-httpd\") pod \"d05ff370-09c4-45c1-b101-a6b277a34d3d\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.786479 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-scripts\") pod \"d05ff370-09c4-45c1-b101-a6b277a34d3d\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.786654 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05ff370-09c4-45c1-b101-a6b277a34d3d-log-httpd\") pod \"d05ff370-09c4-45c1-b101-a6b277a34d3d\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.786790 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-combined-ca-bundle\") pod \"d05ff370-09c4-45c1-b101-a6b277a34d3d\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.786823 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-config-data\") pod \"d05ff370-09c4-45c1-b101-a6b277a34d3d\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.786854 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znq84\" (UniqueName: \"kubernetes.io/projected/d05ff370-09c4-45c1-b101-a6b277a34d3d-kube-api-access-znq84\") pod \"d05ff370-09c4-45c1-b101-a6b277a34d3d\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.786921 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-sg-core-conf-yaml\") pod \"d05ff370-09c4-45c1-b101-a6b277a34d3d\" (UID: \"d05ff370-09c4-45c1-b101-a6b277a34d3d\") " Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.790084 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d05ff370-09c4-45c1-b101-a6b277a34d3d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d05ff370-09c4-45c1-b101-a6b277a34d3d" (UID: "d05ff370-09c4-45c1-b101-a6b277a34d3d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.790456 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d05ff370-09c4-45c1-b101-a6b277a34d3d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d05ff370-09c4-45c1-b101-a6b277a34d3d" (UID: "d05ff370-09c4-45c1-b101-a6b277a34d3d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.798264 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-scripts" (OuterVolumeSpecName: "scripts") pod "d05ff370-09c4-45c1-b101-a6b277a34d3d" (UID: "d05ff370-09c4-45c1-b101-a6b277a34d3d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.809044 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d05ff370-09c4-45c1-b101-a6b277a34d3d-kube-api-access-znq84" (OuterVolumeSpecName: "kube-api-access-znq84") pod "d05ff370-09c4-45c1-b101-a6b277a34d3d" (UID: "d05ff370-09c4-45c1-b101-a6b277a34d3d"). InnerVolumeSpecName "kube-api-access-znq84". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.882545 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d05ff370-09c4-45c1-b101-a6b277a34d3d" (UID: "d05ff370-09c4-45c1-b101-a6b277a34d3d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.889517 4806 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05ff370-09c4-45c1-b101-a6b277a34d3d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.889548 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.889581 4806 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d05ff370-09c4-45c1-b101-a6b277a34d3d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.889593 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znq84\" (UniqueName: \"kubernetes.io/projected/d05ff370-09c4-45c1-b101-a6b277a34d3d-kube-api-access-znq84\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:58 crc kubenswrapper[4806]: I1127 10:38:58.889605 4806 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.030443 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d05ff370-09c4-45c1-b101-a6b277a34d3d" (UID: "d05ff370-09c4-45c1-b101-a6b277a34d3d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.057935 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-config-data" (OuterVolumeSpecName: "config-data") pod "d05ff370-09c4-45c1-b101-a6b277a34d3d" (UID: "d05ff370-09c4-45c1-b101-a6b277a34d3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.092134 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.092319 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05ff370-09c4-45c1-b101-a6b277a34d3d-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.570805 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" event={"ID":"ad03fa3a-e8e2-480d-8980-8a54cd70debb","Type":"ContainerStarted","Data":"7c250018aca01149192de5dd2a29a7b6c84bf13f2f0dfb34806af47c1ce4e8ac"} Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.571199 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" event={"ID":"ad03fa3a-e8e2-480d-8980-8a54cd70debb","Type":"ContainerStarted","Data":"d93600611c7f511ccb277db182ba0d321f06a593676325329a2a2200ffb742a7"} Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.575675 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5bb7f69687-8mg9f" event={"ID":"2a5920ca-ee03-41f8-9886-2c609254e7f7","Type":"ContainerStarted","Data":"457bc93d8154155474097aa84e82e9210a46e3bd039359445b55f819db15b451"} Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.575739 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5bb7f69687-8mg9f" event={"ID":"2a5920ca-ee03-41f8-9886-2c609254e7f7","Type":"ContainerStarted","Data":"3d3036fa33f4b2aaa993f488b0a570535131b82a5b15e932cfa7d49b1d2e365e"} Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.579881 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3f3feff2-b2d0-42b7-9adc-e65686db77f1","Type":"ContainerStarted","Data":"9f729834549912ffac6108781a85bc31b552925be27d2febb0c1b6336148d137"} Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.580046 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3f3feff2-b2d0-42b7-9adc-e65686db77f1" containerName="cinder-api-log" containerID="cri-o://34c085f1489f8bbc3e84ba7457adeea813450610fa09f58cac40df9a2b288c95" gracePeriod=30 Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.580300 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.580333 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="3f3feff2-b2d0-42b7-9adc-e65686db77f1" containerName="cinder-api" containerID="cri-o://9f729834549912ffac6108781a85bc31b552925be27d2febb0c1b6336148d137" gracePeriod=30 Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.593992 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"070121ad-363b-4c59-b6ac-47776d46fb25","Type":"ContainerStarted","Data":"f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271"} Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.600217 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6fb499c486-z7b8m" podStartSLOduration=3.508332546 podStartE2EDuration="7.600199648s" podCreationTimestamp="2025-11-27 10:38:52 +0000 UTC" firstStartedPulling="2025-11-27 10:38:54.095497502 +0000 UTC m=+1038.682088266" lastFinishedPulling="2025-11-27 10:38:58.187364614 +0000 UTC m=+1042.773955368" observedRunningTime="2025-11-27 10:38:59.593565708 +0000 UTC m=+1044.180156472" watchObservedRunningTime="2025-11-27 10:38:59.600199648 +0000 UTC m=+1044.186790412" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.616597 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.616582373 podStartE2EDuration="5.616582373s" podCreationTimestamp="2025-11-27 10:38:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:38:59.613212251 +0000 UTC m=+1044.199803015" watchObservedRunningTime="2025-11-27 10:38:59.616582373 +0000 UTC m=+1044.203173137" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.640023 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5bb7f69687-8mg9f" podStartSLOduration=3.808899556 podStartE2EDuration="7.640005008s" podCreationTimestamp="2025-11-27 10:38:52 +0000 UTC" firstStartedPulling="2025-11-27 10:38:54.352058569 +0000 UTC m=+1038.938649333" lastFinishedPulling="2025-11-27 10:38:58.183164021 +0000 UTC m=+1042.769754785" observedRunningTime="2025-11-27 10:38:59.636077361 +0000 UTC m=+1044.222668135" watchObservedRunningTime="2025-11-27 10:38:59.640005008 +0000 UTC m=+1044.226595772" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.649199 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.651479 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d05ff370-09c4-45c1-b101-a6b277a34d3d","Type":"ContainerDied","Data":"091208a112c811eb72cbf388928ae445197a04ac2d0fd48f1ca61a17ffe2b085"} Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.651554 4806 scope.go:117] "RemoveContainer" containerID="73aeda0d2b661d85a701e6d5162caafce27cfad08c384b94650d9757e3f05fee" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.707174 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.736925 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.761690 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:38:59 crc kubenswrapper[4806]: E1127 10:38:59.762325 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="proxy-httpd" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.762347 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="proxy-httpd" Nov 27 10:38:59 crc kubenswrapper[4806]: E1127 10:38:59.762375 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="ceilometer-central-agent" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.762384 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="ceilometer-central-agent" Nov 27 10:38:59 crc kubenswrapper[4806]: E1127 10:38:59.762417 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="sg-core" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.762425 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="sg-core" Nov 27 10:38:59 crc kubenswrapper[4806]: E1127 10:38:59.762438 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91b270f0-0dec-4c48-ac15-dc748b910066" containerName="init" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.762446 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="91b270f0-0dec-4c48-ac15-dc748b910066" containerName="init" Nov 27 10:38:59 crc kubenswrapper[4806]: E1127 10:38:59.762457 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="ceilometer-notification-agent" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.762465 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="ceilometer-notification-agent" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.762691 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="proxy-httpd" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.762706 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="ceilometer-central-agent" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.762716 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="ceilometer-notification-agent" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.762727 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" containerName="sg-core" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.762738 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="91b270f0-0dec-4c48-ac15-dc748b910066" containerName="init" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.764779 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.778048 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.778903 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.779222 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.822051 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bdefa0d-2c73-43d4-a2e6-759a4f303126-log-httpd\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.822106 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bdefa0d-2c73-43d4-a2e6-759a4f303126-run-httpd\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.822149 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.822175 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkmgh\" (UniqueName: \"kubernetes.io/projected/5bdefa0d-2c73-43d4-a2e6-759a4f303126-kube-api-access-xkmgh\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.822207 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-scripts\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.822255 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.822304 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-config-data\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.852936 4806 scope.go:117] "RemoveContainer" containerID="183d9372b53f6427808ac6d0ce282b15607ed7035e79538be935f9fddf1eb346" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.918487 4806 scope.go:117] "RemoveContainer" containerID="9c03a47efe11a0973c4aaa2985a4039e32eb53c1fe4d3d8393c24291624b7bfb" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.924319 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bdefa0d-2c73-43d4-a2e6-759a4f303126-log-httpd\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.924377 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bdefa0d-2c73-43d4-a2e6-759a4f303126-run-httpd\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.924411 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.924455 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkmgh\" (UniqueName: \"kubernetes.io/projected/5bdefa0d-2c73-43d4-a2e6-759a4f303126-kube-api-access-xkmgh\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.924484 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-scripts\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.924526 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.924566 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-config-data\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.927626 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bdefa0d-2c73-43d4-a2e6-759a4f303126-log-httpd\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.927697 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bdefa0d-2c73-43d4-a2e6-759a4f303126-run-httpd\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.936870 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.938039 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.938451 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-config-data\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.938652 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-scripts\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.958327 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkmgh\" (UniqueName: \"kubernetes.io/projected/5bdefa0d-2c73-43d4-a2e6-759a4f303126-kube-api-access-xkmgh\") pod \"ceilometer-0\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " pod="openstack/ceilometer-0" Nov 27 10:38:59 crc kubenswrapper[4806]: I1127 10:38:59.974507 4806 scope.go:117] "RemoveContainer" containerID="07377df1241e43c09f0052ed513662c35a9c3b54c8608fdc9491c4e98408cdf1" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.122773 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.145056 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d05ff370-09c4-45c1-b101-a6b277a34d3d" path="/var/lib/kubelet/pods/d05ff370-09c4-45c1-b101-a6b277a34d3d/volumes" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.284023 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.413701 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-74c97c7bb-whc9p"] Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.418045 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.432956 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.433810 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.475101 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-combined-ca-bundle\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.475187 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-config-data-custom\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.475264 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25r2f\" (UniqueName: \"kubernetes.io/projected/2f08c589-2ea3-4d79-b068-3c9784701fb0-kube-api-access-25r2f\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.475409 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f08c589-2ea3-4d79-b068-3c9784701fb0-logs\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.475648 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-internal-tls-certs\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.475803 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-config-data\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.475964 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-public-tls-certs\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.491326 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-74c97c7bb-whc9p"] Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.583161 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25r2f\" (UniqueName: \"kubernetes.io/projected/2f08c589-2ea3-4d79-b068-3c9784701fb0-kube-api-access-25r2f\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.583245 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f08c589-2ea3-4d79-b068-3c9784701fb0-logs\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.583303 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-internal-tls-certs\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.583333 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-config-data\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.583368 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-public-tls-certs\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.583416 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-combined-ca-bundle\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.583478 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-config-data-custom\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.593163 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f08c589-2ea3-4d79-b068-3c9784701fb0-logs\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.593854 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-public-tls-certs\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.607825 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-combined-ca-bundle\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.608674 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-internal-tls-certs\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.609749 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-config-data-custom\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.624872 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f08c589-2ea3-4d79-b068-3c9784701fb0-config-data\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.632526 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25r2f\" (UniqueName: \"kubernetes.io/projected/2f08c589-2ea3-4d79-b068-3c9784701fb0-kube-api-access-25r2f\") pod \"barbican-api-74c97c7bb-whc9p\" (UID: \"2f08c589-2ea3-4d79-b068-3c9784701fb0\") " pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.690069 4806 generic.go:334] "Generic (PLEG): container finished" podID="3f3feff2-b2d0-42b7-9adc-e65686db77f1" containerID="34c085f1489f8bbc3e84ba7457adeea813450610fa09f58cac40df9a2b288c95" exitCode=143 Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.690366 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3f3feff2-b2d0-42b7-9adc-e65686db77f1","Type":"ContainerDied","Data":"34c085f1489f8bbc3e84ba7457adeea813450610fa09f58cac40df9a2b288c95"} Nov 27 10:39:00 crc kubenswrapper[4806]: I1127 10:39:00.774431 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:01 crc kubenswrapper[4806]: I1127 10:39:01.209627 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:39:01 crc kubenswrapper[4806]: I1127 10:39:01.582316 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-74c97c7bb-whc9p"] Nov 27 10:39:01 crc kubenswrapper[4806]: W1127 10:39:01.597570 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f08c589_2ea3_4d79_b068_3c9784701fb0.slice/crio-915e942820278f86920ce0d7d153a49227fe221b7f4be2492afba6279eb8eca0 WatchSource:0}: Error finding container 915e942820278f86920ce0d7d153a49227fe221b7f4be2492afba6279eb8eca0: Status 404 returned error can't find the container with id 915e942820278f86920ce0d7d153a49227fe221b7f4be2492afba6279eb8eca0 Nov 27 10:39:01 crc kubenswrapper[4806]: I1127 10:39:01.716881 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bdefa0d-2c73-43d4-a2e6-759a4f303126","Type":"ContainerStarted","Data":"3595307d66db82cdb60fa869361336bbdb7913e58caa9198cc9e9ac3256564cb"} Nov 27 10:39:01 crc kubenswrapper[4806]: I1127 10:39:01.718171 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74c97c7bb-whc9p" event={"ID":"2f08c589-2ea3-4d79-b068-3c9784701fb0","Type":"ContainerStarted","Data":"915e942820278f86920ce0d7d153a49227fe221b7f4be2492afba6279eb8eca0"} Nov 27 10:39:01 crc kubenswrapper[4806]: I1127 10:39:01.722010 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"070121ad-363b-4c59-b6ac-47776d46fb25","Type":"ContainerStarted","Data":"387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90"} Nov 27 10:39:01 crc kubenswrapper[4806]: I1127 10:39:01.756395 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.739440296 podStartE2EDuration="8.756371019s" podCreationTimestamp="2025-11-27 10:38:53 +0000 UTC" firstStartedPulling="2025-11-27 10:38:55.069435903 +0000 UTC m=+1039.656026667" lastFinishedPulling="2025-11-27 10:38:58.086366626 +0000 UTC m=+1042.672957390" observedRunningTime="2025-11-27 10:39:01.749485292 +0000 UTC m=+1046.336076056" watchObservedRunningTime="2025-11-27 10:39:01.756371019 +0000 UTC m=+1046.342961783" Nov 27 10:39:02 crc kubenswrapper[4806]: I1127 10:39:02.743362 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bdefa0d-2c73-43d4-a2e6-759a4f303126","Type":"ContainerStarted","Data":"b7dac30acbd3c50999296d4c8fa332c7cd5b436ddff56febd3195d454ed37fb9"} Nov 27 10:39:02 crc kubenswrapper[4806]: I1127 10:39:02.749478 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74c97c7bb-whc9p" event={"ID":"2f08c589-2ea3-4d79-b068-3c9784701fb0","Type":"ContainerStarted","Data":"92c88a3dfae9377d6cd66a2f6e532439b7a23c55de40bf4f1a972a1f2ef01a26"} Nov 27 10:39:02 crc kubenswrapper[4806]: I1127 10:39:02.749553 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74c97c7bb-whc9p" event={"ID":"2f08c589-2ea3-4d79-b068-3c9784701fb0","Type":"ContainerStarted","Data":"cabd2f97d7ee28ea4ab3f69aa0bfdeaeed77f422b20384670467b89f0ba09902"} Nov 27 10:39:03 crc kubenswrapper[4806]: I1127 10:39:03.767894 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5c76df9f57-fglss" Nov 27 10:39:03 crc kubenswrapper[4806]: I1127 10:39:03.787395 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bdefa0d-2c73-43d4-a2e6-759a4f303126","Type":"ContainerStarted","Data":"e5d9e4f306d95b2f8575ea4c44de2d8830eb1f7e6431cd730b7541aaca49df00"} Nov 27 10:39:03 crc kubenswrapper[4806]: I1127 10:39:03.787472 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:03 crc kubenswrapper[4806]: I1127 10:39:03.787510 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:03 crc kubenswrapper[4806]: I1127 10:39:03.800834 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-74c97c7bb-whc9p" podStartSLOduration=3.80081934 podStartE2EDuration="3.80081934s" podCreationTimestamp="2025-11-27 10:39:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:39:02.784265803 +0000 UTC m=+1047.370856587" watchObservedRunningTime="2025-11-27 10:39:03.80081934 +0000 UTC m=+1048.387410094" Nov 27 10:39:04 crc kubenswrapper[4806]: I1127 10:39:03.968747 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-88c6547cb-ngwks"] Nov 27 10:39:04 crc kubenswrapper[4806]: I1127 10:39:03.969068 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-88c6547cb-ngwks" podUID="c5760001-0bdf-40bb-86ab-34017e1c5297" containerName="neutron-api" containerID="cri-o://3a308f295b808edae84b73c41ec07fe45394d50412d4c94848ea3c1c0df9124e" gracePeriod=30 Nov 27 10:39:04 crc kubenswrapper[4806]: I1127 10:39:03.969430 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-88c6547cb-ngwks" podUID="c5760001-0bdf-40bb-86ab-34017e1c5297" containerName="neutron-httpd" containerID="cri-o://5daf51ee684219f2b33d04b0532759ea8154cf5898ed727c7a505b432ce17808" gracePeriod=30 Nov 27 10:39:04 crc kubenswrapper[4806]: I1127 10:39:04.177655 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 27 10:39:04 crc kubenswrapper[4806]: E1127 10:39:04.360369 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5760001_0bdf_40bb_86ab_34017e1c5297.slice/crio-5daf51ee684219f2b33d04b0532759ea8154cf5898ed727c7a505b432ce17808.scope\": RecentStats: unable to find data in memory cache]" Nov 27 10:39:04 crc kubenswrapper[4806]: I1127 10:39:04.666616 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 27 10:39:04 crc kubenswrapper[4806]: I1127 10:39:04.691431 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:39:04 crc kubenswrapper[4806]: I1127 10:39:04.716535 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9q6vl" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerName="registry-server" probeResult="failure" output=< Nov 27 10:39:04 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 10:39:04 crc kubenswrapper[4806]: > Nov 27 10:39:04 crc kubenswrapper[4806]: I1127 10:39:04.791055 4806 generic.go:334] "Generic (PLEG): container finished" podID="c5760001-0bdf-40bb-86ab-34017e1c5297" containerID="5daf51ee684219f2b33d04b0532759ea8154cf5898ed727c7a505b432ce17808" exitCode=0 Nov 27 10:39:04 crc kubenswrapper[4806]: I1127 10:39:04.791111 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-88c6547cb-ngwks" event={"ID":"c5760001-0bdf-40bb-86ab-34017e1c5297","Type":"ContainerDied","Data":"5daf51ee684219f2b33d04b0532759ea8154cf5898ed727c7a505b432ce17808"} Nov 27 10:39:04 crc kubenswrapper[4806]: I1127 10:39:04.794735 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bdefa0d-2c73-43d4-a2e6-759a4f303126","Type":"ContainerStarted","Data":"a705228ab51536e23620a34956752e1b9931048245734b81637bd4085f66b884"} Nov 27 10:39:04 crc kubenswrapper[4806]: I1127 10:39:04.836394 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-vpvdm"] Nov 27 10:39:04 crc kubenswrapper[4806]: I1127 10:39:04.836625 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" podUID="95cec57d-0ffa-497f-9850-b460bb6a2b53" containerName="dnsmasq-dns" containerID="cri-o://3d49bed70224c73e49e5e3ec59a8006b5eb44a25bc9d05864c7f6d01c7acc41c" gracePeriod=10 Nov 27 10:39:04 crc kubenswrapper[4806]: I1127 10:39:04.932418 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.068111 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" podUID="95cec57d-0ffa-497f-9850-b460bb6a2b53" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.137:5353: connect: connection refused" Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.646808 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.726814 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-ovsdbserver-nb\") pod \"95cec57d-0ffa-497f-9850-b460bb6a2b53\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.727170 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-ovsdbserver-sb\") pod \"95cec57d-0ffa-497f-9850-b460bb6a2b53\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.727397 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hx4c4\" (UniqueName: \"kubernetes.io/projected/95cec57d-0ffa-497f-9850-b460bb6a2b53-kube-api-access-hx4c4\") pod \"95cec57d-0ffa-497f-9850-b460bb6a2b53\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.727427 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-dns-svc\") pod \"95cec57d-0ffa-497f-9850-b460bb6a2b53\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.727472 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-config\") pod \"95cec57d-0ffa-497f-9850-b460bb6a2b53\" (UID: \"95cec57d-0ffa-497f-9850-b460bb6a2b53\") " Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.752411 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95cec57d-0ffa-497f-9850-b460bb6a2b53-kube-api-access-hx4c4" (OuterVolumeSpecName: "kube-api-access-hx4c4") pod "95cec57d-0ffa-497f-9850-b460bb6a2b53" (UID: "95cec57d-0ffa-497f-9850-b460bb6a2b53"). InnerVolumeSpecName "kube-api-access-hx4c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.806953 4806 generic.go:334] "Generic (PLEG): container finished" podID="95cec57d-0ffa-497f-9850-b460bb6a2b53" containerID="3d49bed70224c73e49e5e3ec59a8006b5eb44a25bc9d05864c7f6d01c7acc41c" exitCode=0 Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.808192 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.808276 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" event={"ID":"95cec57d-0ffa-497f-9850-b460bb6a2b53","Type":"ContainerDied","Data":"3d49bed70224c73e49e5e3ec59a8006b5eb44a25bc9d05864c7f6d01c7acc41c"} Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.808339 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-vpvdm" event={"ID":"95cec57d-0ffa-497f-9850-b460bb6a2b53","Type":"ContainerDied","Data":"b719a5c93722e266da4ebe243f4e6f9a11456d9b1d84822f328af6670ee50574"} Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.808358 4806 scope.go:117] "RemoveContainer" containerID="3d49bed70224c73e49e5e3ec59a8006b5eb44a25bc9d05864c7f6d01c7acc41c" Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.808963 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="070121ad-363b-4c59-b6ac-47776d46fb25" containerName="cinder-scheduler" containerID="cri-o://f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271" gracePeriod=30 Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.809047 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="070121ad-363b-4c59-b6ac-47776d46fb25" containerName="probe" containerID="cri-o://387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90" gracePeriod=30 Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.830320 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hx4c4\" (UniqueName: \"kubernetes.io/projected/95cec57d-0ffa-497f-9850-b460bb6a2b53-kube-api-access-hx4c4\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.850121 4806 scope.go:117] "RemoveContainer" containerID="975e1efecc3fc6a8a63c241af367b867208fbb71c479632d8d68dca4dac67a79" Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.871098 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "95cec57d-0ffa-497f-9850-b460bb6a2b53" (UID: "95cec57d-0ffa-497f-9850-b460bb6a2b53"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.929711 4806 scope.go:117] "RemoveContainer" containerID="3d49bed70224c73e49e5e3ec59a8006b5eb44a25bc9d05864c7f6d01c7acc41c" Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.931661 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.932013 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-config" (OuterVolumeSpecName: "config") pod "95cec57d-0ffa-497f-9850-b460bb6a2b53" (UID: "95cec57d-0ffa-497f-9850-b460bb6a2b53"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:39:05 crc kubenswrapper[4806]: E1127 10:39:05.933037 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d49bed70224c73e49e5e3ec59a8006b5eb44a25bc9d05864c7f6d01c7acc41c\": container with ID starting with 3d49bed70224c73e49e5e3ec59a8006b5eb44a25bc9d05864c7f6d01c7acc41c not found: ID does not exist" containerID="3d49bed70224c73e49e5e3ec59a8006b5eb44a25bc9d05864c7f6d01c7acc41c" Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.933101 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d49bed70224c73e49e5e3ec59a8006b5eb44a25bc9d05864c7f6d01c7acc41c"} err="failed to get container status \"3d49bed70224c73e49e5e3ec59a8006b5eb44a25bc9d05864c7f6d01c7acc41c\": rpc error: code = NotFound desc = could not find container \"3d49bed70224c73e49e5e3ec59a8006b5eb44a25bc9d05864c7f6d01c7acc41c\": container with ID starting with 3d49bed70224c73e49e5e3ec59a8006b5eb44a25bc9d05864c7f6d01c7acc41c not found: ID does not exist" Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.933130 4806 scope.go:117] "RemoveContainer" containerID="975e1efecc3fc6a8a63c241af367b867208fbb71c479632d8d68dca4dac67a79" Nov 27 10:39:05 crc kubenswrapper[4806]: E1127 10:39:05.933628 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"975e1efecc3fc6a8a63c241af367b867208fbb71c479632d8d68dca4dac67a79\": container with ID starting with 975e1efecc3fc6a8a63c241af367b867208fbb71c479632d8d68dca4dac67a79 not found: ID does not exist" containerID="975e1efecc3fc6a8a63c241af367b867208fbb71c479632d8d68dca4dac67a79" Nov 27 10:39:05 crc kubenswrapper[4806]: I1127 10:39:05.933667 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"975e1efecc3fc6a8a63c241af367b867208fbb71c479632d8d68dca4dac67a79"} err="failed to get container status \"975e1efecc3fc6a8a63c241af367b867208fbb71c479632d8d68dca4dac67a79\": rpc error: code = NotFound desc = could not find container \"975e1efecc3fc6a8a63c241af367b867208fbb71c479632d8d68dca4dac67a79\": container with ID starting with 975e1efecc3fc6a8a63c241af367b867208fbb71c479632d8d68dca4dac67a79 not found: ID does not exist" Nov 27 10:39:06 crc kubenswrapper[4806]: I1127 10:39:06.003063 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "95cec57d-0ffa-497f-9850-b460bb6a2b53" (UID: "95cec57d-0ffa-497f-9850-b460bb6a2b53"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:39:06 crc kubenswrapper[4806]: I1127 10:39:06.005645 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "95cec57d-0ffa-497f-9850-b460bb6a2b53" (UID: "95cec57d-0ffa-497f-9850-b460bb6a2b53"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:39:06 crc kubenswrapper[4806]: I1127 10:39:06.032828 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:06 crc kubenswrapper[4806]: I1127 10:39:06.032862 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:06 crc kubenswrapper[4806]: I1127 10:39:06.032872 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95cec57d-0ffa-497f-9850-b460bb6a2b53-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:06 crc kubenswrapper[4806]: I1127 10:39:06.198302 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-vpvdm"] Nov 27 10:39:06 crc kubenswrapper[4806]: I1127 10:39:06.276939 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-vpvdm"] Nov 27 10:39:06 crc kubenswrapper[4806]: I1127 10:39:06.826821 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bdefa0d-2c73-43d4-a2e6-759a4f303126","Type":"ContainerStarted","Data":"aa3b570309d6857d45a96eedb84591d5e9fda9bb529f048bb72ca014b87fabcc"} Nov 27 10:39:06 crc kubenswrapper[4806]: I1127 10:39:06.827967 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 10:39:06 crc kubenswrapper[4806]: I1127 10:39:06.863460 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.766884642 podStartE2EDuration="7.863442733s" podCreationTimestamp="2025-11-27 10:38:59 +0000 UTC" firstStartedPulling="2025-11-27 10:39:01.305036789 +0000 UTC m=+1045.891627553" lastFinishedPulling="2025-11-27 10:39:05.40159488 +0000 UTC m=+1049.988185644" observedRunningTime="2025-11-27 10:39:06.86042675 +0000 UTC m=+1051.447017514" watchObservedRunningTime="2025-11-27 10:39:06.863442733 +0000 UTC m=+1051.450033507" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.474664 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.636042 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/070121ad-363b-4c59-b6ac-47776d46fb25-etc-machine-id\") pod \"070121ad-363b-4c59-b6ac-47776d46fb25\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.636148 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-config-data\") pod \"070121ad-363b-4c59-b6ac-47776d46fb25\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.636189 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-combined-ca-bundle\") pod \"070121ad-363b-4c59-b6ac-47776d46fb25\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.636291 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-config-data-custom\") pod \"070121ad-363b-4c59-b6ac-47776d46fb25\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.636310 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-scripts\") pod \"070121ad-363b-4c59-b6ac-47776d46fb25\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.636403 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb4xs\" (UniqueName: \"kubernetes.io/projected/070121ad-363b-4c59-b6ac-47776d46fb25-kube-api-access-rb4xs\") pod \"070121ad-363b-4c59-b6ac-47776d46fb25\" (UID: \"070121ad-363b-4c59-b6ac-47776d46fb25\") " Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.637287 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/070121ad-363b-4c59-b6ac-47776d46fb25-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "070121ad-363b-4c59-b6ac-47776d46fb25" (UID: "070121ad-363b-4c59-b6ac-47776d46fb25"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.653562 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.664585 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/070121ad-363b-4c59-b6ac-47776d46fb25-kube-api-access-rb4xs" (OuterVolumeSpecName: "kube-api-access-rb4xs") pod "070121ad-363b-4c59-b6ac-47776d46fb25" (UID: "070121ad-363b-4c59-b6ac-47776d46fb25"). InnerVolumeSpecName "kube-api-access-rb4xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.672404 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "070121ad-363b-4c59-b6ac-47776d46fb25" (UID: "070121ad-363b-4c59-b6ac-47776d46fb25"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.673965 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-scripts" (OuterVolumeSpecName: "scripts") pod "070121ad-363b-4c59-b6ac-47776d46fb25" (UID: "070121ad-363b-4c59-b6ac-47776d46fb25"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.742829 4806 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/070121ad-363b-4c59-b6ac-47776d46fb25-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.742911 4806 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.742927 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.742939 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb4xs\" (UniqueName: \"kubernetes.io/projected/070121ad-363b-4c59-b6ac-47776d46fb25-kube-api-access-rb4xs\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.787448 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "070121ad-363b-4c59-b6ac-47776d46fb25" (UID: "070121ad-363b-4c59-b6ac-47776d46fb25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.844452 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.854495 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-config-data" (OuterVolumeSpecName: "config-data") pod "070121ad-363b-4c59-b6ac-47776d46fb25" (UID: "070121ad-363b-4c59-b6ac-47776d46fb25"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.863455 4806 generic.go:334] "Generic (PLEG): container finished" podID="070121ad-363b-4c59-b6ac-47776d46fb25" containerID="387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90" exitCode=0 Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.863494 4806 generic.go:334] "Generic (PLEG): container finished" podID="070121ad-363b-4c59-b6ac-47776d46fb25" containerID="f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271" exitCode=0 Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.863623 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"070121ad-363b-4c59-b6ac-47776d46fb25","Type":"ContainerDied","Data":"387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90"} Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.863626 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.863658 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"070121ad-363b-4c59-b6ac-47776d46fb25","Type":"ContainerDied","Data":"f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271"} Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.863682 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"070121ad-363b-4c59-b6ac-47776d46fb25","Type":"ContainerDied","Data":"001528992b8a04707b3c48dc9f0efb0de30217fe0bfe41f8d302412e1ed8befc"} Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.863701 4806 scope.go:117] "RemoveContainer" containerID="387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.892778 4806 generic.go:334] "Generic (PLEG): container finished" podID="c5760001-0bdf-40bb-86ab-34017e1c5297" containerID="3a308f295b808edae84b73c41ec07fe45394d50412d4c94848ea3c1c0df9124e" exitCode=0 Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.892913 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-88c6547cb-ngwks" event={"ID":"c5760001-0bdf-40bb-86ab-34017e1c5297","Type":"ContainerDied","Data":"3a308f295b808edae84b73c41ec07fe45394d50412d4c94848ea3c1c0df9124e"} Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.915399 4806 scope.go:117] "RemoveContainer" containerID="f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.946407 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070121ad-363b-4c59-b6ac-47776d46fb25-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.985367 4806 scope.go:117] "RemoveContainer" containerID="387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90" Nov 27 10:39:07 crc kubenswrapper[4806]: E1127 10:39:07.997842 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90\": container with ID starting with 387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90 not found: ID does not exist" containerID="387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.997887 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90"} err="failed to get container status \"387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90\": rpc error: code = NotFound desc = could not find container \"387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90\": container with ID starting with 387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90 not found: ID does not exist" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.997920 4806 scope.go:117] "RemoveContainer" containerID="f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271" Nov 27 10:39:07 crc kubenswrapper[4806]: I1127 10:39:07.998000 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 10:39:08 crc kubenswrapper[4806]: E1127 10:39:08.007836 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271\": container with ID starting with f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271 not found: ID does not exist" containerID="f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.007888 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271"} err="failed to get container status \"f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271\": rpc error: code = NotFound desc = could not find container \"f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271\": container with ID starting with f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271 not found: ID does not exist" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.007916 4806 scope.go:117] "RemoveContainer" containerID="387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.008263 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90"} err="failed to get container status \"387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90\": rpc error: code = NotFound desc = could not find container \"387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90\": container with ID starting with 387b45bd7c8029a47f25f6b6ab0ebe35fb792472e2cd6f0f5a90078682f2dd90 not found: ID does not exist" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.008319 4806 scope.go:117] "RemoveContainer" containerID="f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.008537 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271"} err="failed to get container status \"f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271\": rpc error: code = NotFound desc = could not find container \"f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271\": container with ID starting with f8e25c216b995f67da3d0c6469a3183a521158d22709a82ead9d6b7862e50271 not found: ID does not exist" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.043361 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.049603 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 10:39:08 crc kubenswrapper[4806]: E1127 10:39:08.050893 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070121ad-363b-4c59-b6ac-47776d46fb25" containerName="probe" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.050913 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="070121ad-363b-4c59-b6ac-47776d46fb25" containerName="probe" Nov 27 10:39:08 crc kubenswrapper[4806]: E1127 10:39:08.050924 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95cec57d-0ffa-497f-9850-b460bb6a2b53" containerName="dnsmasq-dns" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.050931 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="95cec57d-0ffa-497f-9850-b460bb6a2b53" containerName="dnsmasq-dns" Nov 27 10:39:08 crc kubenswrapper[4806]: E1127 10:39:08.050941 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95cec57d-0ffa-497f-9850-b460bb6a2b53" containerName="init" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.050947 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="95cec57d-0ffa-497f-9850-b460bb6a2b53" containerName="init" Nov 27 10:39:08 crc kubenswrapper[4806]: E1127 10:39:08.050961 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070121ad-363b-4c59-b6ac-47776d46fb25" containerName="cinder-scheduler" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.050968 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="070121ad-363b-4c59-b6ac-47776d46fb25" containerName="cinder-scheduler" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.052633 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="95cec57d-0ffa-497f-9850-b460bb6a2b53" containerName="dnsmasq-dns" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.052668 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="070121ad-363b-4c59-b6ac-47776d46fb25" containerName="cinder-scheduler" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.052685 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="070121ad-363b-4c59-b6ac-47776d46fb25" containerName="probe" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.057410 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.058073 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.061211 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.094611 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.172500 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="070121ad-363b-4c59-b6ac-47776d46fb25" path="/var/lib/kubelet/pods/070121ad-363b-4c59-b6ac-47776d46fb25/volumes" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.173418 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95cec57d-0ffa-497f-9850-b460bb6a2b53" path="/var/lib/kubelet/pods/95cec57d-0ffa-497f-9850-b460bb6a2b53/volumes" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.256665 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zms72\" (UniqueName: \"kubernetes.io/projected/c5760001-0bdf-40bb-86ab-34017e1c5297-kube-api-access-zms72\") pod \"c5760001-0bdf-40bb-86ab-34017e1c5297\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.256729 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-httpd-config\") pod \"c5760001-0bdf-40bb-86ab-34017e1c5297\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.256789 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-config\") pod \"c5760001-0bdf-40bb-86ab-34017e1c5297\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.256833 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-ovndb-tls-certs\") pod \"c5760001-0bdf-40bb-86ab-34017e1c5297\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.256873 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-combined-ca-bundle\") pod \"c5760001-0bdf-40bb-86ab-34017e1c5297\" (UID: \"c5760001-0bdf-40bb-86ab-34017e1c5297\") " Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.257081 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-config-data\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.257152 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.257181 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-scripts\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.257218 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.257254 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdr2k\" (UniqueName: \"kubernetes.io/projected/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-kube-api-access-kdr2k\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.257290 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.269755 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5760001-0bdf-40bb-86ab-34017e1c5297-kube-api-access-zms72" (OuterVolumeSpecName: "kube-api-access-zms72") pod "c5760001-0bdf-40bb-86ab-34017e1c5297" (UID: "c5760001-0bdf-40bb-86ab-34017e1c5297"). InnerVolumeSpecName "kube-api-access-zms72". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.276586 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "c5760001-0bdf-40bb-86ab-34017e1c5297" (UID: "c5760001-0bdf-40bb-86ab-34017e1c5297"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.326162 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5760001-0bdf-40bb-86ab-34017e1c5297" (UID: "c5760001-0bdf-40bb-86ab-34017e1c5297"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.335424 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-config" (OuterVolumeSpecName: "config") pod "c5760001-0bdf-40bb-86ab-34017e1c5297" (UID: "c5760001-0bdf-40bb-86ab-34017e1c5297"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.358997 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.359052 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-scripts\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.359087 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.359104 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdr2k\" (UniqueName: \"kubernetes.io/projected/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-kube-api-access-kdr2k\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.359144 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.359186 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-config-data\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.359249 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.359260 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.359270 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zms72\" (UniqueName: \"kubernetes.io/projected/c5760001-0bdf-40bb-86ab-34017e1c5297-kube-api-access-zms72\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.359278 4806 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.361124 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.363995 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.373114 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-config-data\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.379852 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.380329 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-scripts\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.383083 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdr2k\" (UniqueName: \"kubernetes.io/projected/845fe2ce-0faf-4c9a-b1f3-ce642c58a316-kube-api-access-kdr2k\") pod \"cinder-scheduler-0\" (UID: \"845fe2ce-0faf-4c9a-b1f3-ce642c58a316\") " pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.403379 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "c5760001-0bdf-40bb-86ab-34017e1c5297" (UID: "c5760001-0bdf-40bb-86ab-34017e1c5297"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.403797 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.460639 4806 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5760001-0bdf-40bb-86ab-34017e1c5297-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.491568 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8ccc54b-kzmfd" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.145:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.500859 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.957720 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-88c6547cb-ngwks" event={"ID":"c5760001-0bdf-40bb-86ab-34017e1c5297","Type":"ContainerDied","Data":"9dd285faca5c35f034597e2edf7906617b876286ddc2744e6bdeff2986f64781"} Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.958296 4806 scope.go:117] "RemoveContainer" containerID="5daf51ee684219f2b33d04b0532759ea8154cf5898ed727c7a505b432ce17808" Nov 27 10:39:08 crc kubenswrapper[4806]: I1127 10:39:08.958470 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-88c6547cb-ngwks" Nov 27 10:39:09 crc kubenswrapper[4806]: I1127 10:39:09.065455 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-88c6547cb-ngwks"] Nov 27 10:39:09 crc kubenswrapper[4806]: I1127 10:39:09.181398 4806 scope.go:117] "RemoveContainer" containerID="3a308f295b808edae84b73c41ec07fe45394d50412d4c94848ea3c1c0df9124e" Nov 27 10:39:09 crc kubenswrapper[4806]: I1127 10:39:09.226407 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-88c6547cb-ngwks"] Nov 27 10:39:09 crc kubenswrapper[4806]: I1127 10:39:09.252923 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 27 10:39:09 crc kubenswrapper[4806]: I1127 10:39:09.821661 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="3f3feff2-b2d0-42b7-9adc-e65686db77f1" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.148:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 10:39:10 crc kubenswrapper[4806]: I1127 10:39:10.018208 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"845fe2ce-0faf-4c9a-b1f3-ce642c58a316","Type":"ContainerStarted","Data":"88b66c9238e78fed97f0a5bfd490ecc33332efd8d7c33aab3c39b62aee092e4d"} Nov 27 10:39:10 crc kubenswrapper[4806]: I1127 10:39:10.127035 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5760001-0bdf-40bb-86ab-34017e1c5297" path="/var/lib/kubelet/pods/c5760001-0bdf-40bb-86ab-34017e1c5297/volumes" Nov 27 10:39:11 crc kubenswrapper[4806]: I1127 10:39:11.034359 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"845fe2ce-0faf-4c9a-b1f3-ce642c58a316","Type":"ContainerStarted","Data":"dc74c5579a995f6c6263b1a01883d193b86f7ddea4d60a0af68b14040c860d30"} Nov 27 10:39:11 crc kubenswrapper[4806]: I1127 10:39:11.034918 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"845fe2ce-0faf-4c9a-b1f3-ce642c58a316","Type":"ContainerStarted","Data":"d2e5aa541bd92b1e3a28281c1d3d563bc4b0ad190a146cae84759a032ad09c6d"} Nov 27 10:39:11 crc kubenswrapper[4806]: I1127 10:39:11.067495 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.067475126 podStartE2EDuration="4.067475126s" podCreationTimestamp="2025-11-27 10:39:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:39:11.056500959 +0000 UTC m=+1055.643091723" watchObservedRunningTime="2025-11-27 10:39:11.067475126 +0000 UTC m=+1055.654065890" Nov 27 10:39:11 crc kubenswrapper[4806]: I1127 10:39:11.421521 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-8ccc54b-kzmfd" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.145:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 10:39:11 crc kubenswrapper[4806]: I1127 10:39:11.824545 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5689867474-qpmnv" Nov 27 10:39:12 crc kubenswrapper[4806]: I1127 10:39:12.151388 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5689867474-qpmnv" Nov 27 10:39:13 crc kubenswrapper[4806]: I1127 10:39:13.416675 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 27 10:39:13 crc kubenswrapper[4806]: I1127 10:39:13.492510 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8ccc54b-kzmfd" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.145:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 10:39:13 crc kubenswrapper[4806]: I1127 10:39:13.753731 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-c9d8f7ccf-5z2jj" Nov 27 10:39:14 crc kubenswrapper[4806]: I1127 10:39:14.450196 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:39:14 crc kubenswrapper[4806]: I1127 10:39:14.450266 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:39:14 crc kubenswrapper[4806]: I1127 10:39:14.629670 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 27 10:39:14 crc kubenswrapper[4806]: I1127 10:39:14.703332 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9q6vl" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerName="registry-server" probeResult="failure" output=< Nov 27 10:39:14 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 10:39:14 crc kubenswrapper[4806]: > Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.112525 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.632589 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-74c97c7bb-whc9p" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.723756 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8ccc54b-kzmfd"] Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.723974 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8ccc54b-kzmfd" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerName="barbican-api-log" containerID="cri-o://6e514a446147a41fe562cb099f03806ccf4f5dbd165ea55189506f356e536b89" gracePeriod=30 Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.724394 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8ccc54b-kzmfd" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerName="barbican-api" containerID="cri-o://e933eeb67d80d9dbf4b5a96b9e24bfb6ca8410b0b02cb6ba88b3f0b57ad11d52" gracePeriod=30 Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.911916 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 27 10:39:15 crc kubenswrapper[4806]: E1127 10:39:15.912538 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5760001-0bdf-40bb-86ab-34017e1c5297" containerName="neutron-httpd" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.912553 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5760001-0bdf-40bb-86ab-34017e1c5297" containerName="neutron-httpd" Nov 27 10:39:15 crc kubenswrapper[4806]: E1127 10:39:15.912582 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5760001-0bdf-40bb-86ab-34017e1c5297" containerName="neutron-api" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.912591 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5760001-0bdf-40bb-86ab-34017e1c5297" containerName="neutron-api" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.912737 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5760001-0bdf-40bb-86ab-34017e1c5297" containerName="neutron-api" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.912760 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5760001-0bdf-40bb-86ab-34017e1c5297" containerName="neutron-httpd" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.913280 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.934295 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-bld8v" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.935246 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.935543 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.948264 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.963337 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01bb064b-2da1-4695-8e6e-8c635ff6dbaa-combined-ca-bundle\") pod \"openstackclient\" (UID: \"01bb064b-2da1-4695-8e6e-8c635ff6dbaa\") " pod="openstack/openstackclient" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.963449 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdgkj\" (UniqueName: \"kubernetes.io/projected/01bb064b-2da1-4695-8e6e-8c635ff6dbaa-kube-api-access-fdgkj\") pod \"openstackclient\" (UID: \"01bb064b-2da1-4695-8e6e-8c635ff6dbaa\") " pod="openstack/openstackclient" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.963512 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/01bb064b-2da1-4695-8e6e-8c635ff6dbaa-openstack-config-secret\") pod \"openstackclient\" (UID: \"01bb064b-2da1-4695-8e6e-8c635ff6dbaa\") " pod="openstack/openstackclient" Nov 27 10:39:15 crc kubenswrapper[4806]: I1127 10:39:15.963547 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/01bb064b-2da1-4695-8e6e-8c635ff6dbaa-openstack-config\") pod \"openstackclient\" (UID: \"01bb064b-2da1-4695-8e6e-8c635ff6dbaa\") " pod="openstack/openstackclient" Nov 27 10:39:16 crc kubenswrapper[4806]: I1127 10:39:16.065334 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/01bb064b-2da1-4695-8e6e-8c635ff6dbaa-openstack-config\") pod \"openstackclient\" (UID: \"01bb064b-2da1-4695-8e6e-8c635ff6dbaa\") " pod="openstack/openstackclient" Nov 27 10:39:16 crc kubenswrapper[4806]: I1127 10:39:16.065459 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01bb064b-2da1-4695-8e6e-8c635ff6dbaa-combined-ca-bundle\") pod \"openstackclient\" (UID: \"01bb064b-2da1-4695-8e6e-8c635ff6dbaa\") " pod="openstack/openstackclient" Nov 27 10:39:16 crc kubenswrapper[4806]: I1127 10:39:16.065490 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdgkj\" (UniqueName: \"kubernetes.io/projected/01bb064b-2da1-4695-8e6e-8c635ff6dbaa-kube-api-access-fdgkj\") pod \"openstackclient\" (UID: \"01bb064b-2da1-4695-8e6e-8c635ff6dbaa\") " pod="openstack/openstackclient" Nov 27 10:39:16 crc kubenswrapper[4806]: I1127 10:39:16.065533 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/01bb064b-2da1-4695-8e6e-8c635ff6dbaa-openstack-config-secret\") pod \"openstackclient\" (UID: \"01bb064b-2da1-4695-8e6e-8c635ff6dbaa\") " pod="openstack/openstackclient" Nov 27 10:39:16 crc kubenswrapper[4806]: I1127 10:39:16.078292 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/01bb064b-2da1-4695-8e6e-8c635ff6dbaa-openstack-config\") pod \"openstackclient\" (UID: \"01bb064b-2da1-4695-8e6e-8c635ff6dbaa\") " pod="openstack/openstackclient" Nov 27 10:39:16 crc kubenswrapper[4806]: I1127 10:39:16.091545 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/01bb064b-2da1-4695-8e6e-8c635ff6dbaa-openstack-config-secret\") pod \"openstackclient\" (UID: \"01bb064b-2da1-4695-8e6e-8c635ff6dbaa\") " pod="openstack/openstackclient" Nov 27 10:39:16 crc kubenswrapper[4806]: I1127 10:39:16.093089 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01bb064b-2da1-4695-8e6e-8c635ff6dbaa-combined-ca-bundle\") pod \"openstackclient\" (UID: \"01bb064b-2da1-4695-8e6e-8c635ff6dbaa\") " pod="openstack/openstackclient" Nov 27 10:39:16 crc kubenswrapper[4806]: I1127 10:39:16.100913 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdgkj\" (UniqueName: \"kubernetes.io/projected/01bb064b-2da1-4695-8e6e-8c635ff6dbaa-kube-api-access-fdgkj\") pod \"openstackclient\" (UID: \"01bb064b-2da1-4695-8e6e-8c635ff6dbaa\") " pod="openstack/openstackclient" Nov 27 10:39:16 crc kubenswrapper[4806]: I1127 10:39:16.133532 4806 generic.go:334] "Generic (PLEG): container finished" podID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerID="6e514a446147a41fe562cb099f03806ccf4f5dbd165ea55189506f356e536b89" exitCode=143 Nov 27 10:39:16 crc kubenswrapper[4806]: I1127 10:39:16.143200 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8ccc54b-kzmfd" event={"ID":"3967a997-3385-4163-97e8-fe3ef5eaf57c","Type":"ContainerDied","Data":"6e514a446147a41fe562cb099f03806ccf4f5dbd165ea55189506f356e536b89"} Nov 27 10:39:16 crc kubenswrapper[4806]: I1127 10:39:16.251767 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 27 10:39:16 crc kubenswrapper[4806]: I1127 10:39:16.904630 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 27 10:39:17 crc kubenswrapper[4806]: I1127 10:39:17.144595 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"01bb064b-2da1-4695-8e6e-8c635ff6dbaa","Type":"ContainerStarted","Data":"e878ba646c08f096974ff4d7e4de6aa3ec1bf2bd40add46c4872371fb212682d"} Nov 27 10:39:18 crc kubenswrapper[4806]: I1127 10:39:18.833854 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 27 10:39:19 crc kubenswrapper[4806]: I1127 10:39:19.479775 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8ccc54b-kzmfd" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.145:9311/healthcheck\": read tcp 10.217.0.2:60642->10.217.0.145:9311: read: connection reset by peer" Nov 27 10:39:19 crc kubenswrapper[4806]: I1127 10:39:19.479773 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8ccc54b-kzmfd" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.145:9311/healthcheck\": read tcp 10.217.0.2:60632->10.217.0.145:9311: read: connection reset by peer" Nov 27 10:39:19 crc kubenswrapper[4806]: I1127 10:39:19.965774 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.054275 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3967a997-3385-4163-97e8-fe3ef5eaf57c-logs\") pod \"3967a997-3385-4163-97e8-fe3ef5eaf57c\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.054462 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-combined-ca-bundle\") pod \"3967a997-3385-4163-97e8-fe3ef5eaf57c\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.054514 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbm4t\" (UniqueName: \"kubernetes.io/projected/3967a997-3385-4163-97e8-fe3ef5eaf57c-kube-api-access-xbm4t\") pod \"3967a997-3385-4163-97e8-fe3ef5eaf57c\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.054545 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-config-data-custom\") pod \"3967a997-3385-4163-97e8-fe3ef5eaf57c\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.054566 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-config-data\") pod \"3967a997-3385-4163-97e8-fe3ef5eaf57c\" (UID: \"3967a997-3385-4163-97e8-fe3ef5eaf57c\") " Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.059095 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3967a997-3385-4163-97e8-fe3ef5eaf57c-logs" (OuterVolumeSpecName: "logs") pod "3967a997-3385-4163-97e8-fe3ef5eaf57c" (UID: "3967a997-3385-4163-97e8-fe3ef5eaf57c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.069466 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3967a997-3385-4163-97e8-fe3ef5eaf57c-kube-api-access-xbm4t" (OuterVolumeSpecName: "kube-api-access-xbm4t") pod "3967a997-3385-4163-97e8-fe3ef5eaf57c" (UID: "3967a997-3385-4163-97e8-fe3ef5eaf57c"). InnerVolumeSpecName "kube-api-access-xbm4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.083470 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3967a997-3385-4163-97e8-fe3ef5eaf57c" (UID: "3967a997-3385-4163-97e8-fe3ef5eaf57c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.153426 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3967a997-3385-4163-97e8-fe3ef5eaf57c" (UID: "3967a997-3385-4163-97e8-fe3ef5eaf57c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.156608 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.156635 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbm4t\" (UniqueName: \"kubernetes.io/projected/3967a997-3385-4163-97e8-fe3ef5eaf57c-kube-api-access-xbm4t\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.156646 4806 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.156656 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3967a997-3385-4163-97e8-fe3ef5eaf57c-logs\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.169185 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-config-data" (OuterVolumeSpecName: "config-data") pod "3967a997-3385-4163-97e8-fe3ef5eaf57c" (UID: "3967a997-3385-4163-97e8-fe3ef5eaf57c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.192493 4806 generic.go:334] "Generic (PLEG): container finished" podID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerID="e933eeb67d80d9dbf4b5a96b9e24bfb6ca8410b0b02cb6ba88b3f0b57ad11d52" exitCode=0 Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.192739 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8ccc54b-kzmfd" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.192726 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8ccc54b-kzmfd" event={"ID":"3967a997-3385-4163-97e8-fe3ef5eaf57c","Type":"ContainerDied","Data":"e933eeb67d80d9dbf4b5a96b9e24bfb6ca8410b0b02cb6ba88b3f0b57ad11d52"} Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.193159 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8ccc54b-kzmfd" event={"ID":"3967a997-3385-4163-97e8-fe3ef5eaf57c","Type":"ContainerDied","Data":"c4ca8251cf5920a77392245356caa8224b5dfd9d0ca3fd3da897f2961c6334d3"} Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.193183 4806 scope.go:117] "RemoveContainer" containerID="e933eeb67d80d9dbf4b5a96b9e24bfb6ca8410b0b02cb6ba88b3f0b57ad11d52" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.230055 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8ccc54b-kzmfd"] Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.239568 4806 scope.go:117] "RemoveContainer" containerID="6e514a446147a41fe562cb099f03806ccf4f5dbd165ea55189506f356e536b89" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.252444 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-8ccc54b-kzmfd"] Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.258632 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3967a997-3385-4163-97e8-fe3ef5eaf57c-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.273446 4806 scope.go:117] "RemoveContainer" containerID="e933eeb67d80d9dbf4b5a96b9e24bfb6ca8410b0b02cb6ba88b3f0b57ad11d52" Nov 27 10:39:20 crc kubenswrapper[4806]: E1127 10:39:20.275254 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e933eeb67d80d9dbf4b5a96b9e24bfb6ca8410b0b02cb6ba88b3f0b57ad11d52\": container with ID starting with e933eeb67d80d9dbf4b5a96b9e24bfb6ca8410b0b02cb6ba88b3f0b57ad11d52 not found: ID does not exist" containerID="e933eeb67d80d9dbf4b5a96b9e24bfb6ca8410b0b02cb6ba88b3f0b57ad11d52" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.275306 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e933eeb67d80d9dbf4b5a96b9e24bfb6ca8410b0b02cb6ba88b3f0b57ad11d52"} err="failed to get container status \"e933eeb67d80d9dbf4b5a96b9e24bfb6ca8410b0b02cb6ba88b3f0b57ad11d52\": rpc error: code = NotFound desc = could not find container \"e933eeb67d80d9dbf4b5a96b9e24bfb6ca8410b0b02cb6ba88b3f0b57ad11d52\": container with ID starting with e933eeb67d80d9dbf4b5a96b9e24bfb6ca8410b0b02cb6ba88b3f0b57ad11d52 not found: ID does not exist" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.275338 4806 scope.go:117] "RemoveContainer" containerID="6e514a446147a41fe562cb099f03806ccf4f5dbd165ea55189506f356e536b89" Nov 27 10:39:20 crc kubenswrapper[4806]: E1127 10:39:20.275792 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e514a446147a41fe562cb099f03806ccf4f5dbd165ea55189506f356e536b89\": container with ID starting with 6e514a446147a41fe562cb099f03806ccf4f5dbd165ea55189506f356e536b89 not found: ID does not exist" containerID="6e514a446147a41fe562cb099f03806ccf4f5dbd165ea55189506f356e536b89" Nov 27 10:39:20 crc kubenswrapper[4806]: I1127 10:39:20.275822 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e514a446147a41fe562cb099f03806ccf4f5dbd165ea55189506f356e536b89"} err="failed to get container status \"6e514a446147a41fe562cb099f03806ccf4f5dbd165ea55189506f356e536b89\": rpc error: code = NotFound desc = could not find container \"6e514a446147a41fe562cb099f03806ccf4f5dbd165ea55189506f356e536b89\": container with ID starting with 6e514a446147a41fe562cb099f03806ccf4f5dbd165ea55189506f356e536b89 not found: ID does not exist" Nov 27 10:39:22 crc kubenswrapper[4806]: I1127 10:39:22.126340 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" path="/var/lib/kubelet/pods/3967a997-3385-4163-97e8-fe3ef5eaf57c/volumes" Nov 27 10:39:24 crc kubenswrapper[4806]: I1127 10:39:24.722001 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9q6vl" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerName="registry-server" probeResult="failure" output=< Nov 27 10:39:24 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 10:39:24 crc kubenswrapper[4806]: > Nov 27 10:39:27 crc kubenswrapper[4806]: I1127 10:39:27.860634 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:39:27 crc kubenswrapper[4806]: I1127 10:39:27.862057 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="sg-core" containerID="cri-o://a705228ab51536e23620a34956752e1b9931048245734b81637bd4085f66b884" gracePeriod=30 Nov 27 10:39:27 crc kubenswrapper[4806]: I1127 10:39:27.862105 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="ceilometer-notification-agent" containerID="cri-o://e5d9e4f306d95b2f8575ea4c44de2d8830eb1f7e6431cd730b7541aaca49df00" gracePeriod=30 Nov 27 10:39:27 crc kubenswrapper[4806]: I1127 10:39:27.862116 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="proxy-httpd" containerID="cri-o://aa3b570309d6857d45a96eedb84591d5e9fda9bb529f048bb72ca014b87fabcc" gracePeriod=30 Nov 27 10:39:27 crc kubenswrapper[4806]: I1127 10:39:27.862005 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="ceilometer-central-agent" containerID="cri-o://b7dac30acbd3c50999296d4c8fa332c7cd5b436ddff56febd3195d454ed37fb9" gracePeriod=30 Nov 27 10:39:27 crc kubenswrapper[4806]: I1127 10:39:27.874212 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.149:3000/\": EOF" Nov 27 10:39:28 crc kubenswrapper[4806]: I1127 10:39:28.280624 4806 generic.go:334] "Generic (PLEG): container finished" podID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerID="aa3b570309d6857d45a96eedb84591d5e9fda9bb529f048bb72ca014b87fabcc" exitCode=0 Nov 27 10:39:28 crc kubenswrapper[4806]: I1127 10:39:28.280879 4806 generic.go:334] "Generic (PLEG): container finished" podID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerID="a705228ab51536e23620a34956752e1b9931048245734b81637bd4085f66b884" exitCode=2 Nov 27 10:39:28 crc kubenswrapper[4806]: I1127 10:39:28.280902 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bdefa0d-2c73-43d4-a2e6-759a4f303126","Type":"ContainerDied","Data":"aa3b570309d6857d45a96eedb84591d5e9fda9bb529f048bb72ca014b87fabcc"} Nov 27 10:39:28 crc kubenswrapper[4806]: I1127 10:39:28.280929 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bdefa0d-2c73-43d4-a2e6-759a4f303126","Type":"ContainerDied","Data":"a705228ab51536e23620a34956752e1b9931048245734b81637bd4085f66b884"} Nov 27 10:39:29 crc kubenswrapper[4806]: I1127 10:39:29.296439 4806 generic.go:334] "Generic (PLEG): container finished" podID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerID="e5d9e4f306d95b2f8575ea4c44de2d8830eb1f7e6431cd730b7541aaca49df00" exitCode=0 Nov 27 10:39:29 crc kubenswrapper[4806]: I1127 10:39:29.296485 4806 generic.go:334] "Generic (PLEG): container finished" podID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerID="b7dac30acbd3c50999296d4c8fa332c7cd5b436ddff56febd3195d454ed37fb9" exitCode=0 Nov 27 10:39:29 crc kubenswrapper[4806]: I1127 10:39:29.296516 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bdefa0d-2c73-43d4-a2e6-759a4f303126","Type":"ContainerDied","Data":"e5d9e4f306d95b2f8575ea4c44de2d8830eb1f7e6431cd730b7541aaca49df00"} Nov 27 10:39:29 crc kubenswrapper[4806]: I1127 10:39:29.296574 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bdefa0d-2c73-43d4-a2e6-759a4f303126","Type":"ContainerDied","Data":"b7dac30acbd3c50999296d4c8fa332c7cd5b436ddff56febd3195d454ed37fb9"} Nov 27 10:39:29 crc kubenswrapper[4806]: I1127 10:39:29.775573 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="3f3feff2-b2d0-42b7-9adc-e65686db77f1" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.148:8776/healthcheck\": dial tcp 10.217.0.148:8776: connect: connection refused" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.124378 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.149:3000/\": dial tcp 10.217.0.149:3000: connect: connection refused" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.311467 4806 generic.go:334] "Generic (PLEG): container finished" podID="3f3feff2-b2d0-42b7-9adc-e65686db77f1" containerID="9f729834549912ffac6108781a85bc31b552925be27d2febb0c1b6336148d137" exitCode=137 Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.311537 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3f3feff2-b2d0-42b7-9adc-e65686db77f1","Type":"ContainerDied","Data":"9f729834549912ffac6108781a85bc31b552925be27d2febb0c1b6336148d137"} Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.669090 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.709519 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.756945 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-scripts\") pod \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.756995 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f3feff2-b2d0-42b7-9adc-e65686db77f1-logs\") pod \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.757079 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-combined-ca-bundle\") pod \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.757129 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-config-data\") pod \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.757181 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f3feff2-b2d0-42b7-9adc-e65686db77f1-etc-machine-id\") pod \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.757224 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-config-data-custom\") pod \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.757350 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2fcf\" (UniqueName: \"kubernetes.io/projected/3f3feff2-b2d0-42b7-9adc-e65686db77f1-kube-api-access-g2fcf\") pod \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\" (UID: \"3f3feff2-b2d0-42b7-9adc-e65686db77f1\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.771366 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f3feff2-b2d0-42b7-9adc-e65686db77f1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3f3feff2-b2d0-42b7-9adc-e65686db77f1" (UID: "3f3feff2-b2d0-42b7-9adc-e65686db77f1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.772651 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f3feff2-b2d0-42b7-9adc-e65686db77f1-kube-api-access-g2fcf" (OuterVolumeSpecName: "kube-api-access-g2fcf") pod "3f3feff2-b2d0-42b7-9adc-e65686db77f1" (UID: "3f3feff2-b2d0-42b7-9adc-e65686db77f1"). InnerVolumeSpecName "kube-api-access-g2fcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.772965 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f3feff2-b2d0-42b7-9adc-e65686db77f1-logs" (OuterVolumeSpecName: "logs") pod "3f3feff2-b2d0-42b7-9adc-e65686db77f1" (UID: "3f3feff2-b2d0-42b7-9adc-e65686db77f1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.773943 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-scripts" (OuterVolumeSpecName: "scripts") pod "3f3feff2-b2d0-42b7-9adc-e65686db77f1" (UID: "3f3feff2-b2d0-42b7-9adc-e65686db77f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.775158 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3f3feff2-b2d0-42b7-9adc-e65686db77f1" (UID: "3f3feff2-b2d0-42b7-9adc-e65686db77f1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.805137 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f3feff2-b2d0-42b7-9adc-e65686db77f1" (UID: "3f3feff2-b2d0-42b7-9adc-e65686db77f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.832457 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-config-data" (OuterVolumeSpecName: "config-data") pod "3f3feff2-b2d0-42b7-9adc-e65686db77f1" (UID: "3f3feff2-b2d0-42b7-9adc-e65686db77f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.858937 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-config-data\") pod \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.859020 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkmgh\" (UniqueName: \"kubernetes.io/projected/5bdefa0d-2c73-43d4-a2e6-759a4f303126-kube-api-access-xkmgh\") pod \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.859073 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-scripts\") pod \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.859110 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-sg-core-conf-yaml\") pod \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.859130 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bdefa0d-2c73-43d4-a2e6-759a4f303126-log-httpd\") pod \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.859213 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-combined-ca-bundle\") pod \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.859246 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bdefa0d-2c73-43d4-a2e6-759a4f303126-run-httpd\") pod \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\" (UID: \"5bdefa0d-2c73-43d4-a2e6-759a4f303126\") " Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.859600 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.859618 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.859627 4806 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f3feff2-b2d0-42b7-9adc-e65686db77f1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.859639 4806 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.859650 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2fcf\" (UniqueName: \"kubernetes.io/projected/3f3feff2-b2d0-42b7-9adc-e65686db77f1-kube-api-access-g2fcf\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.859662 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f3feff2-b2d0-42b7-9adc-e65686db77f1-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.859671 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f3feff2-b2d0-42b7-9adc-e65686db77f1-logs\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.860003 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bdefa0d-2c73-43d4-a2e6-759a4f303126-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5bdefa0d-2c73-43d4-a2e6-759a4f303126" (UID: "5bdefa0d-2c73-43d4-a2e6-759a4f303126"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.860203 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bdefa0d-2c73-43d4-a2e6-759a4f303126-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5bdefa0d-2c73-43d4-a2e6-759a4f303126" (UID: "5bdefa0d-2c73-43d4-a2e6-759a4f303126"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.865163 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bdefa0d-2c73-43d4-a2e6-759a4f303126-kube-api-access-xkmgh" (OuterVolumeSpecName: "kube-api-access-xkmgh") pod "5bdefa0d-2c73-43d4-a2e6-759a4f303126" (UID: "5bdefa0d-2c73-43d4-a2e6-759a4f303126"). InnerVolumeSpecName "kube-api-access-xkmgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.868407 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-scripts" (OuterVolumeSpecName: "scripts") pod "5bdefa0d-2c73-43d4-a2e6-759a4f303126" (UID: "5bdefa0d-2c73-43d4-a2e6-759a4f303126"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.904779 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5bdefa0d-2c73-43d4-a2e6-759a4f303126" (UID: "5bdefa0d-2c73-43d4-a2e6-759a4f303126"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.962141 4806 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bdefa0d-2c73-43d4-a2e6-759a4f303126-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.962177 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkmgh\" (UniqueName: \"kubernetes.io/projected/5bdefa0d-2c73-43d4-a2e6-759a4f303126-kube-api-access-xkmgh\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.962188 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.962196 4806 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.962203 4806 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5bdefa0d-2c73-43d4-a2e6-759a4f303126-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.968399 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bdefa0d-2c73-43d4-a2e6-759a4f303126" (UID: "5bdefa0d-2c73-43d4-a2e6-759a4f303126"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:30 crc kubenswrapper[4806]: I1127 10:39:30.973758 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-config-data" (OuterVolumeSpecName: "config-data") pod "5bdefa0d-2c73-43d4-a2e6-759a4f303126" (UID: "5bdefa0d-2c73-43d4-a2e6-759a4f303126"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.077999 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.078049 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bdefa0d-2c73-43d4-a2e6-759a4f303126-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.321583 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5bdefa0d-2c73-43d4-a2e6-759a4f303126","Type":"ContainerDied","Data":"3595307d66db82cdb60fa869361336bbdb7913e58caa9198cc9e9ac3256564cb"} Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.321660 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.321716 4806 scope.go:117] "RemoveContainer" containerID="aa3b570309d6857d45a96eedb84591d5e9fda9bb529f048bb72ca014b87fabcc" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.323337 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"01bb064b-2da1-4695-8e6e-8c635ff6dbaa","Type":"ContainerStarted","Data":"628612daefbef08323d0474b2f9638fdc72d108199c0ec641f5a7a9f46249dc3"} Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.325195 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3f3feff2-b2d0-42b7-9adc-e65686db77f1","Type":"ContainerDied","Data":"135864be863d704c327f5aec71c6b78303cfe46461261a72e5e41562d08b6122"} Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.325275 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.342152 4806 scope.go:117] "RemoveContainer" containerID="a705228ab51536e23620a34956752e1b9931048245734b81637bd4085f66b884" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.349274 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.902894809 podStartE2EDuration="16.349252186s" podCreationTimestamp="2025-11-27 10:39:15 +0000 UTC" firstStartedPulling="2025-11-27 10:39:16.931690432 +0000 UTC m=+1061.518281196" lastFinishedPulling="2025-11-27 10:39:30.378047789 +0000 UTC m=+1074.964638573" observedRunningTime="2025-11-27 10:39:31.345932256 +0000 UTC m=+1075.932523020" watchObservedRunningTime="2025-11-27 10:39:31.349252186 +0000 UTC m=+1075.935842950" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.375555 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.381592 4806 scope.go:117] "RemoveContainer" containerID="e5d9e4f306d95b2f8575ea4c44de2d8830eb1f7e6431cd730b7541aaca49df00" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.389775 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.401245 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.413848 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.415842 4806 scope.go:117] "RemoveContainer" containerID="b7dac30acbd3c50999296d4c8fa332c7cd5b436ddff56febd3195d454ed37fb9" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.421949 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 27 10:39:31 crc kubenswrapper[4806]: E1127 10:39:31.422422 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerName="barbican-api" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422441 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerName="barbican-api" Nov 27 10:39:31 crc kubenswrapper[4806]: E1127 10:39:31.422459 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f3feff2-b2d0-42b7-9adc-e65686db77f1" containerName="cinder-api" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422465 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f3feff2-b2d0-42b7-9adc-e65686db77f1" containerName="cinder-api" Nov 27 10:39:31 crc kubenswrapper[4806]: E1127 10:39:31.422482 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f3feff2-b2d0-42b7-9adc-e65686db77f1" containerName="cinder-api-log" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422489 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f3feff2-b2d0-42b7-9adc-e65686db77f1" containerName="cinder-api-log" Nov 27 10:39:31 crc kubenswrapper[4806]: E1127 10:39:31.422505 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="proxy-httpd" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422511 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="proxy-httpd" Nov 27 10:39:31 crc kubenswrapper[4806]: E1127 10:39:31.422525 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerName="barbican-api-log" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422532 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerName="barbican-api-log" Nov 27 10:39:31 crc kubenswrapper[4806]: E1127 10:39:31.422546 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="ceilometer-notification-agent" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422553 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="ceilometer-notification-agent" Nov 27 10:39:31 crc kubenswrapper[4806]: E1127 10:39:31.422562 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="sg-core" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422568 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="sg-core" Nov 27 10:39:31 crc kubenswrapper[4806]: E1127 10:39:31.422580 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="ceilometer-central-agent" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422586 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="ceilometer-central-agent" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422741 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f3feff2-b2d0-42b7-9adc-e65686db77f1" containerName="cinder-api" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422751 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="sg-core" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422762 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerName="barbican-api" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422772 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="ceilometer-central-agent" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422780 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3967a997-3385-4163-97e8-fe3ef5eaf57c" containerName="barbican-api-log" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422789 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="proxy-httpd" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422799 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f3feff2-b2d0-42b7-9adc-e65686db77f1" containerName="cinder-api-log" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.422809 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" containerName="ceilometer-notification-agent" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.434071 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.434192 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.436923 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.437143 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.447024 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.447120 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.455547 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.457858 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.457992 4806 scope.go:117] "RemoveContainer" containerID="9f729834549912ffac6108781a85bc31b552925be27d2febb0c1b6336148d137" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.458048 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.462156 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.501674 4806 scope.go:117] "RemoveContainer" containerID="34c085f1489f8bbc3e84ba7457adeea813450610fa09f58cac40df9a2b288c95" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588493 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588552 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588589 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-config-data-custom\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588641 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-scripts\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588664 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-config-data\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588680 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e91180ab-be63-4bdd-87f7-9a9975f35634-logs\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588713 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4091a1a3-33e0-42ac-935a-a96b43770c62-log-httpd\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588732 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588767 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588793 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e91180ab-be63-4bdd-87f7-9a9975f35634-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588819 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-scripts\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588845 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4091a1a3-33e0-42ac-935a-a96b43770c62-run-httpd\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588869 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb5x8\" (UniqueName: \"kubernetes.io/projected/4091a1a3-33e0-42ac-935a-a96b43770c62-kube-api-access-xb5x8\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588896 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-config-data\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588915 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcgsr\" (UniqueName: \"kubernetes.io/projected/e91180ab-be63-4bdd-87f7-9a9975f35634-kube-api-access-jcgsr\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.588938 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691312 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-scripts\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691382 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-config-data\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691410 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e91180ab-be63-4bdd-87f7-9a9975f35634-logs\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691442 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4091a1a3-33e0-42ac-935a-a96b43770c62-log-httpd\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691459 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691490 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691513 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e91180ab-be63-4bdd-87f7-9a9975f35634-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691546 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-scripts\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691565 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4091a1a3-33e0-42ac-935a-a96b43770c62-run-httpd\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691588 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb5x8\" (UniqueName: \"kubernetes.io/projected/4091a1a3-33e0-42ac-935a-a96b43770c62-kube-api-access-xb5x8\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691611 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-config-data\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691627 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcgsr\" (UniqueName: \"kubernetes.io/projected/e91180ab-be63-4bdd-87f7-9a9975f35634-kube-api-access-jcgsr\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691650 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691679 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691711 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.691750 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-config-data-custom\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.695738 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4091a1a3-33e0-42ac-935a-a96b43770c62-log-httpd\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.696025 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e91180ab-be63-4bdd-87f7-9a9975f35634-logs\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.696707 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4091a1a3-33e0-42ac-935a-a96b43770c62-run-httpd\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.701981 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-config-data-custom\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.704767 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-scripts\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.704863 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e91180ab-be63-4bdd-87f7-9a9975f35634-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.708273 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-scripts\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.709292 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.709933 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.710462 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.712025 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.728855 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcgsr\" (UniqueName: \"kubernetes.io/projected/e91180ab-be63-4bdd-87f7-9a9975f35634-kube-api-access-jcgsr\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.729702 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb5x8\" (UniqueName: \"kubernetes.io/projected/4091a1a3-33e0-42ac-935a-a96b43770c62-kube-api-access-xb5x8\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.730020 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.732786 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e91180ab-be63-4bdd-87f7-9a9975f35634-config-data\") pod \"cinder-api-0\" (UID: \"e91180ab-be63-4bdd-87f7-9a9975f35634\") " pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.733728 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-config-data\") pod \"ceilometer-0\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " pod="openstack/ceilometer-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.775616 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 27 10:39:31 crc kubenswrapper[4806]: I1127 10:39:31.789262 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:39:32 crc kubenswrapper[4806]: I1127 10:39:32.142364 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f3feff2-b2d0-42b7-9adc-e65686db77f1" path="/var/lib/kubelet/pods/3f3feff2-b2d0-42b7-9adc-e65686db77f1/volumes" Nov 27 10:39:32 crc kubenswrapper[4806]: I1127 10:39:32.144688 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bdefa0d-2c73-43d4-a2e6-759a4f303126" path="/var/lib/kubelet/pods/5bdefa0d-2c73-43d4-a2e6-759a4f303126/volumes" Nov 27 10:39:32 crc kubenswrapper[4806]: I1127 10:39:32.286007 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 27 10:39:32 crc kubenswrapper[4806]: W1127 10:39:32.296176 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode91180ab_be63_4bdd_87f7_9a9975f35634.slice/crio-52f4d8aadb767fee07a9240873fcaf5622113fecaa65e9f6eee2169798577640 WatchSource:0}: Error finding container 52f4d8aadb767fee07a9240873fcaf5622113fecaa65e9f6eee2169798577640: Status 404 returned error can't find the container with id 52f4d8aadb767fee07a9240873fcaf5622113fecaa65e9f6eee2169798577640 Nov 27 10:39:32 crc kubenswrapper[4806]: I1127 10:39:32.349811 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e91180ab-be63-4bdd-87f7-9a9975f35634","Type":"ContainerStarted","Data":"52f4d8aadb767fee07a9240873fcaf5622113fecaa65e9f6eee2169798577640"} Nov 27 10:39:32 crc kubenswrapper[4806]: I1127 10:39:32.387329 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:39:32 crc kubenswrapper[4806]: W1127 10:39:32.391325 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4091a1a3_33e0_42ac_935a_a96b43770c62.slice/crio-e75e6688f07e54c2a84da523caaa690dbbd9e1a7424fa935677f27b1115faf57 WatchSource:0}: Error finding container e75e6688f07e54c2a84da523caaa690dbbd9e1a7424fa935677f27b1115faf57: Status 404 returned error can't find the container with id e75e6688f07e54c2a84da523caaa690dbbd9e1a7424fa935677f27b1115faf57 Nov 27 10:39:33 crc kubenswrapper[4806]: I1127 10:39:33.404521 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e91180ab-be63-4bdd-87f7-9a9975f35634","Type":"ContainerStarted","Data":"f8bfb6b42be57fc949242e78a8ab892494db6678d4ad8360909cdac4ae36f2bb"} Nov 27 10:39:33 crc kubenswrapper[4806]: I1127 10:39:33.406805 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4091a1a3-33e0-42ac-935a-a96b43770c62","Type":"ContainerStarted","Data":"899dcf3d182a47a7127b251270a2b7d444b62796bbcf8c943b8c4eb9f9afb835"} Nov 27 10:39:33 crc kubenswrapper[4806]: I1127 10:39:33.406867 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4091a1a3-33e0-42ac-935a-a96b43770c62","Type":"ContainerStarted","Data":"e75e6688f07e54c2a84da523caaa690dbbd9e1a7424fa935677f27b1115faf57"} Nov 27 10:39:33 crc kubenswrapper[4806]: I1127 10:39:33.720537 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:39:33 crc kubenswrapper[4806]: I1127 10:39:33.842538 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:39:33 crc kubenswrapper[4806]: I1127 10:39:33.971966 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9q6vl"] Nov 27 10:39:34 crc kubenswrapper[4806]: I1127 10:39:34.109908 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:39:34 crc kubenswrapper[4806]: I1127 10:39:34.416779 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e91180ab-be63-4bdd-87f7-9a9975f35634","Type":"ContainerStarted","Data":"3687a95326c5465c4aa731967264771ab47231342e85923e1c8bba3a920bcfa5"} Nov 27 10:39:34 crc kubenswrapper[4806]: I1127 10:39:34.416910 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 27 10:39:34 crc kubenswrapper[4806]: I1127 10:39:34.418772 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4091a1a3-33e0-42ac-935a-a96b43770c62","Type":"ContainerStarted","Data":"08c19d395ef8d5c9a9fecf6d0ef362b2afdc4edac7489f8230730d4c22fbb5b9"} Nov 27 10:39:34 crc kubenswrapper[4806]: I1127 10:39:34.438655 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.438638064 podStartE2EDuration="3.438638064s" podCreationTimestamp="2025-11-27 10:39:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:39:34.433078633 +0000 UTC m=+1079.019669397" watchObservedRunningTime="2025-11-27 10:39:34.438638064 +0000 UTC m=+1079.025228828" Nov 27 10:39:35 crc kubenswrapper[4806]: I1127 10:39:35.432834 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4091a1a3-33e0-42ac-935a-a96b43770c62","Type":"ContainerStarted","Data":"e9da1debb9c953660dd0afee85de3504760da50ecc3dbaeee4041c629f8e3c47"} Nov 27 10:39:35 crc kubenswrapper[4806]: I1127 10:39:35.433180 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9q6vl" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerName="registry-server" containerID="cri-o://b3cd2c17c4874ce0e2745f24437a88a7f004d53c35a7ef2f510e9466cb07dd1a" gracePeriod=2 Nov 27 10:39:35 crc kubenswrapper[4806]: I1127 10:39:35.950283 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.097151 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b50a58-b14e-4ae3-ae29-98e5181aa828-catalog-content\") pod \"95b50a58-b14e-4ae3-ae29-98e5181aa828\" (UID: \"95b50a58-b14e-4ae3-ae29-98e5181aa828\") " Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.097324 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b50a58-b14e-4ae3-ae29-98e5181aa828-utilities\") pod \"95b50a58-b14e-4ae3-ae29-98e5181aa828\" (UID: \"95b50a58-b14e-4ae3-ae29-98e5181aa828\") " Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.097430 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvjnz\" (UniqueName: \"kubernetes.io/projected/95b50a58-b14e-4ae3-ae29-98e5181aa828-kube-api-access-pvjnz\") pod \"95b50a58-b14e-4ae3-ae29-98e5181aa828\" (UID: \"95b50a58-b14e-4ae3-ae29-98e5181aa828\") " Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.097733 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95b50a58-b14e-4ae3-ae29-98e5181aa828-utilities" (OuterVolumeSpecName: "utilities") pod "95b50a58-b14e-4ae3-ae29-98e5181aa828" (UID: "95b50a58-b14e-4ae3-ae29-98e5181aa828"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.120208 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95b50a58-b14e-4ae3-ae29-98e5181aa828-kube-api-access-pvjnz" (OuterVolumeSpecName: "kube-api-access-pvjnz") pod "95b50a58-b14e-4ae3-ae29-98e5181aa828" (UID: "95b50a58-b14e-4ae3-ae29-98e5181aa828"). InnerVolumeSpecName "kube-api-access-pvjnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.214418 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95b50a58-b14e-4ae3-ae29-98e5181aa828-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.214451 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvjnz\" (UniqueName: \"kubernetes.io/projected/95b50a58-b14e-4ae3-ae29-98e5181aa828-kube-api-access-pvjnz\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.297715 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95b50a58-b14e-4ae3-ae29-98e5181aa828-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95b50a58-b14e-4ae3-ae29-98e5181aa828" (UID: "95b50a58-b14e-4ae3-ae29-98e5181aa828"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.333530 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95b50a58-b14e-4ae3-ae29-98e5181aa828-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.443968 4806 generic.go:334] "Generic (PLEG): container finished" podID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerID="b3cd2c17c4874ce0e2745f24437a88a7f004d53c35a7ef2f510e9466cb07dd1a" exitCode=0 Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.444180 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9q6vl" event={"ID":"95b50a58-b14e-4ae3-ae29-98e5181aa828","Type":"ContainerDied","Data":"b3cd2c17c4874ce0e2745f24437a88a7f004d53c35a7ef2f510e9466cb07dd1a"} Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.444326 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9q6vl" event={"ID":"95b50a58-b14e-4ae3-ae29-98e5181aa828","Type":"ContainerDied","Data":"522117fb9f162931de795793793aa168b4ac8f138689e77fedf4291c96570e3f"} Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.444335 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9q6vl" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.444349 4806 scope.go:117] "RemoveContainer" containerID="b3cd2c17c4874ce0e2745f24437a88a7f004d53c35a7ef2f510e9466cb07dd1a" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.449758 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4091a1a3-33e0-42ac-935a-a96b43770c62","Type":"ContainerStarted","Data":"4d5346db599f0475d4af9dc97a5f424b6acd4a7fddbb47f122d99bc78f6a2e31"} Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.449905 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="ceilometer-central-agent" containerID="cri-o://899dcf3d182a47a7127b251270a2b7d444b62796bbcf8c943b8c4eb9f9afb835" gracePeriod=30 Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.450130 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.450376 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="proxy-httpd" containerID="cri-o://4d5346db599f0475d4af9dc97a5f424b6acd4a7fddbb47f122d99bc78f6a2e31" gracePeriod=30 Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.450430 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="sg-core" containerID="cri-o://e9da1debb9c953660dd0afee85de3504760da50ecc3dbaeee4041c629f8e3c47" gracePeriod=30 Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.450465 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="ceilometer-notification-agent" containerID="cri-o://08c19d395ef8d5c9a9fecf6d0ef362b2afdc4edac7489f8230730d4c22fbb5b9" gracePeriod=30 Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.476002 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.677976017 podStartE2EDuration="5.475982322s" podCreationTimestamp="2025-11-27 10:39:31 +0000 UTC" firstStartedPulling="2025-11-27 10:39:32.396151186 +0000 UTC m=+1076.982741950" lastFinishedPulling="2025-11-27 10:39:36.194157491 +0000 UTC m=+1080.780748255" observedRunningTime="2025-11-27 10:39:36.475668434 +0000 UTC m=+1081.062259198" watchObservedRunningTime="2025-11-27 10:39:36.475982322 +0000 UTC m=+1081.062573086" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.478690 4806 scope.go:117] "RemoveContainer" containerID="b8b296b8d2494494a350a87831356c852babf046758d6744b025bcd0010c9fa2" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.512648 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9q6vl"] Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.513453 4806 scope.go:117] "RemoveContainer" containerID="9eab18251d6748b7bdfacf0f098d4a00a5a39e3890097d25103566c1e599117f" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.521391 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9q6vl"] Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.542727 4806 scope.go:117] "RemoveContainer" containerID="b3cd2c17c4874ce0e2745f24437a88a7f004d53c35a7ef2f510e9466cb07dd1a" Nov 27 10:39:36 crc kubenswrapper[4806]: E1127 10:39:36.543085 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3cd2c17c4874ce0e2745f24437a88a7f004d53c35a7ef2f510e9466cb07dd1a\": container with ID starting with b3cd2c17c4874ce0e2745f24437a88a7f004d53c35a7ef2f510e9466cb07dd1a not found: ID does not exist" containerID="b3cd2c17c4874ce0e2745f24437a88a7f004d53c35a7ef2f510e9466cb07dd1a" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.543119 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3cd2c17c4874ce0e2745f24437a88a7f004d53c35a7ef2f510e9466cb07dd1a"} err="failed to get container status \"b3cd2c17c4874ce0e2745f24437a88a7f004d53c35a7ef2f510e9466cb07dd1a\": rpc error: code = NotFound desc = could not find container \"b3cd2c17c4874ce0e2745f24437a88a7f004d53c35a7ef2f510e9466cb07dd1a\": container with ID starting with b3cd2c17c4874ce0e2745f24437a88a7f004d53c35a7ef2f510e9466cb07dd1a not found: ID does not exist" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.543141 4806 scope.go:117] "RemoveContainer" containerID="b8b296b8d2494494a350a87831356c852babf046758d6744b025bcd0010c9fa2" Nov 27 10:39:36 crc kubenswrapper[4806]: E1127 10:39:36.543335 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8b296b8d2494494a350a87831356c852babf046758d6744b025bcd0010c9fa2\": container with ID starting with b8b296b8d2494494a350a87831356c852babf046758d6744b025bcd0010c9fa2 not found: ID does not exist" containerID="b8b296b8d2494494a350a87831356c852babf046758d6744b025bcd0010c9fa2" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.543361 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8b296b8d2494494a350a87831356c852babf046758d6744b025bcd0010c9fa2"} err="failed to get container status \"b8b296b8d2494494a350a87831356c852babf046758d6744b025bcd0010c9fa2\": rpc error: code = NotFound desc = could not find container \"b8b296b8d2494494a350a87831356c852babf046758d6744b025bcd0010c9fa2\": container with ID starting with b8b296b8d2494494a350a87831356c852babf046758d6744b025bcd0010c9fa2 not found: ID does not exist" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.543377 4806 scope.go:117] "RemoveContainer" containerID="9eab18251d6748b7bdfacf0f098d4a00a5a39e3890097d25103566c1e599117f" Nov 27 10:39:36 crc kubenswrapper[4806]: E1127 10:39:36.543990 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eab18251d6748b7bdfacf0f098d4a00a5a39e3890097d25103566c1e599117f\": container with ID starting with 9eab18251d6748b7bdfacf0f098d4a00a5a39e3890097d25103566c1e599117f not found: ID does not exist" containerID="9eab18251d6748b7bdfacf0f098d4a00a5a39e3890097d25103566c1e599117f" Nov 27 10:39:36 crc kubenswrapper[4806]: I1127 10:39:36.544010 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eab18251d6748b7bdfacf0f098d4a00a5a39e3890097d25103566c1e599117f"} err="failed to get container status \"9eab18251d6748b7bdfacf0f098d4a00a5a39e3890097d25103566c1e599117f\": rpc error: code = NotFound desc = could not find container \"9eab18251d6748b7bdfacf0f098d4a00a5a39e3890097d25103566c1e599117f\": container with ID starting with 9eab18251d6748b7bdfacf0f098d4a00a5a39e3890097d25103566c1e599117f not found: ID does not exist" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.459770 4806 generic.go:334] "Generic (PLEG): container finished" podID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerID="e9da1debb9c953660dd0afee85de3504760da50ecc3dbaeee4041c629f8e3c47" exitCode=2 Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.459816 4806 generic.go:334] "Generic (PLEG): container finished" podID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerID="08c19d395ef8d5c9a9fecf6d0ef362b2afdc4edac7489f8230730d4c22fbb5b9" exitCode=0 Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.459818 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4091a1a3-33e0-42ac-935a-a96b43770c62","Type":"ContainerDied","Data":"e9da1debb9c953660dd0afee85de3504760da50ecc3dbaeee4041c629f8e3c47"} Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.459857 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4091a1a3-33e0-42ac-935a-a96b43770c62","Type":"ContainerDied","Data":"08c19d395ef8d5c9a9fecf6d0ef362b2afdc4edac7489f8230730d4c22fbb5b9"} Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.651518 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-28sdq"] Nov 27 10:39:37 crc kubenswrapper[4806]: E1127 10:39:37.652257 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerName="extract-content" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.652281 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerName="extract-content" Nov 27 10:39:37 crc kubenswrapper[4806]: E1127 10:39:37.652317 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerName="registry-server" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.652323 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerName="registry-server" Nov 27 10:39:37 crc kubenswrapper[4806]: E1127 10:39:37.652332 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerName="extract-utilities" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.652339 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerName="extract-utilities" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.652506 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" containerName="registry-server" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.653097 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-28sdq" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.680005 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-28sdq"] Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.751116 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-4fbch"] Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.752108 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4fbch" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.753309 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7a0c344-9e24-43af-badd-b761a5005c56-operator-scripts\") pod \"nova-api-db-create-28sdq\" (UID: \"e7a0c344-9e24-43af-badd-b761a5005c56\") " pod="openstack/nova-api-db-create-28sdq" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.753347 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2n9t\" (UniqueName: \"kubernetes.io/projected/e7a0c344-9e24-43af-badd-b761a5005c56-kube-api-access-z2n9t\") pod \"nova-api-db-create-28sdq\" (UID: \"e7a0c344-9e24-43af-badd-b761a5005c56\") " pod="openstack/nova-api-db-create-28sdq" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.764695 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4fbch"] Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.851758 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-7795-account-create-update-wrpdm"] Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.852735 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7795-account-create-update-wrpdm" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.854905 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.854961 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7a0c344-9e24-43af-badd-b761a5005c56-operator-scripts\") pod \"nova-api-db-create-28sdq\" (UID: \"e7a0c344-9e24-43af-badd-b761a5005c56\") " pod="openstack/nova-api-db-create-28sdq" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.855010 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2n9t\" (UniqueName: \"kubernetes.io/projected/e7a0c344-9e24-43af-badd-b761a5005c56-kube-api-access-z2n9t\") pod \"nova-api-db-create-28sdq\" (UID: \"e7a0c344-9e24-43af-badd-b761a5005c56\") " pod="openstack/nova-api-db-create-28sdq" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.855045 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7l6h\" (UniqueName: \"kubernetes.io/projected/727574b8-48dd-4ca6-882f-e84934bf1240-kube-api-access-t7l6h\") pod \"nova-cell0-db-create-4fbch\" (UID: \"727574b8-48dd-4ca6-882f-e84934bf1240\") " pod="openstack/nova-cell0-db-create-4fbch" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.855073 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/727574b8-48dd-4ca6-882f-e84934bf1240-operator-scripts\") pod \"nova-cell0-db-create-4fbch\" (UID: \"727574b8-48dd-4ca6-882f-e84934bf1240\") " pod="openstack/nova-cell0-db-create-4fbch" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.856173 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7a0c344-9e24-43af-badd-b761a5005c56-operator-scripts\") pod \"nova-api-db-create-28sdq\" (UID: \"e7a0c344-9e24-43af-badd-b761a5005c56\") " pod="openstack/nova-api-db-create-28sdq" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.864514 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7795-account-create-update-wrpdm"] Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.889450 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2n9t\" (UniqueName: \"kubernetes.io/projected/e7a0c344-9e24-43af-badd-b761a5005c56-kube-api-access-z2n9t\") pod \"nova-api-db-create-28sdq\" (UID: \"e7a0c344-9e24-43af-badd-b761a5005c56\") " pod="openstack/nova-api-db-create-28sdq" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.950258 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-fjg6t"] Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.955701 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fjg6t" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.956499 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wglg\" (UniqueName: \"kubernetes.io/projected/f9041e87-4dd0-442d-a056-59d77e5134d0-kube-api-access-9wglg\") pod \"nova-api-7795-account-create-update-wrpdm\" (UID: \"f9041e87-4dd0-442d-a056-59d77e5134d0\") " pod="openstack/nova-api-7795-account-create-update-wrpdm" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.956609 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7l6h\" (UniqueName: \"kubernetes.io/projected/727574b8-48dd-4ca6-882f-e84934bf1240-kube-api-access-t7l6h\") pod \"nova-cell0-db-create-4fbch\" (UID: \"727574b8-48dd-4ca6-882f-e84934bf1240\") " pod="openstack/nova-cell0-db-create-4fbch" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.956641 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/727574b8-48dd-4ca6-882f-e84934bf1240-operator-scripts\") pod \"nova-cell0-db-create-4fbch\" (UID: \"727574b8-48dd-4ca6-882f-e84934bf1240\") " pod="openstack/nova-cell0-db-create-4fbch" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.956686 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9041e87-4dd0-442d-a056-59d77e5134d0-operator-scripts\") pod \"nova-api-7795-account-create-update-wrpdm\" (UID: \"f9041e87-4dd0-442d-a056-59d77e5134d0\") " pod="openstack/nova-api-7795-account-create-update-wrpdm" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.957886 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/727574b8-48dd-4ca6-882f-e84934bf1240-operator-scripts\") pod \"nova-cell0-db-create-4fbch\" (UID: \"727574b8-48dd-4ca6-882f-e84934bf1240\") " pod="openstack/nova-cell0-db-create-4fbch" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.969797 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fjg6t"] Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.976273 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7l6h\" (UniqueName: \"kubernetes.io/projected/727574b8-48dd-4ca6-882f-e84934bf1240-kube-api-access-t7l6h\") pod \"nova-cell0-db-create-4fbch\" (UID: \"727574b8-48dd-4ca6-882f-e84934bf1240\") " pod="openstack/nova-cell0-db-create-4fbch" Nov 27 10:39:37 crc kubenswrapper[4806]: I1127 10:39:37.986661 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-28sdq" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.058779 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3954c8e-9c73-4bb3-9db6-e64c3a86c480-operator-scripts\") pod \"nova-cell1-db-create-fjg6t\" (UID: \"d3954c8e-9c73-4bb3-9db6-e64c3a86c480\") " pod="openstack/nova-cell1-db-create-fjg6t" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.058885 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9041e87-4dd0-442d-a056-59d77e5134d0-operator-scripts\") pod \"nova-api-7795-account-create-update-wrpdm\" (UID: \"f9041e87-4dd0-442d-a056-59d77e5134d0\") " pod="openstack/nova-api-7795-account-create-update-wrpdm" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.058917 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q49fx\" (UniqueName: \"kubernetes.io/projected/d3954c8e-9c73-4bb3-9db6-e64c3a86c480-kube-api-access-q49fx\") pod \"nova-cell1-db-create-fjg6t\" (UID: \"d3954c8e-9c73-4bb3-9db6-e64c3a86c480\") " pod="openstack/nova-cell1-db-create-fjg6t" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.058960 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wglg\" (UniqueName: \"kubernetes.io/projected/f9041e87-4dd0-442d-a056-59d77e5134d0-kube-api-access-9wglg\") pod \"nova-api-7795-account-create-update-wrpdm\" (UID: \"f9041e87-4dd0-442d-a056-59d77e5134d0\") " pod="openstack/nova-api-7795-account-create-update-wrpdm" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.059885 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9041e87-4dd0-442d-a056-59d77e5134d0-operator-scripts\") pod \"nova-api-7795-account-create-update-wrpdm\" (UID: \"f9041e87-4dd0-442d-a056-59d77e5134d0\") " pod="openstack/nova-api-7795-account-create-update-wrpdm" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.077525 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4fbch" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.089154 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wglg\" (UniqueName: \"kubernetes.io/projected/f9041e87-4dd0-442d-a056-59d77e5134d0-kube-api-access-9wglg\") pod \"nova-api-7795-account-create-update-wrpdm\" (UID: \"f9041e87-4dd0-442d-a056-59d77e5134d0\") " pod="openstack/nova-api-7795-account-create-update-wrpdm" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.093594 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-4bc5-account-create-update-d58fg"] Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.095849 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.110634 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.168084 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q49fx\" (UniqueName: \"kubernetes.io/projected/d3954c8e-9c73-4bb3-9db6-e64c3a86c480-kube-api-access-q49fx\") pod \"nova-cell1-db-create-fjg6t\" (UID: \"d3954c8e-9c73-4bb3-9db6-e64c3a86c480\") " pod="openstack/nova-cell1-db-create-fjg6t" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.169000 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3954c8e-9c73-4bb3-9db6-e64c3a86c480-operator-scripts\") pod \"nova-cell1-db-create-fjg6t\" (UID: \"d3954c8e-9c73-4bb3-9db6-e64c3a86c480\") " pod="openstack/nova-cell1-db-create-fjg6t" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.169882 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3954c8e-9c73-4bb3-9db6-e64c3a86c480-operator-scripts\") pod \"nova-cell1-db-create-fjg6t\" (UID: \"d3954c8e-9c73-4bb3-9db6-e64c3a86c480\") " pod="openstack/nova-cell1-db-create-fjg6t" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.179396 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7795-account-create-update-wrpdm" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.181015 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95b50a58-b14e-4ae3-ae29-98e5181aa828" path="/var/lib/kubelet/pods/95b50a58-b14e-4ae3-ae29-98e5181aa828/volumes" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.184150 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-4bc5-account-create-update-d58fg"] Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.198357 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q49fx\" (UniqueName: \"kubernetes.io/projected/d3954c8e-9c73-4bb3-9db6-e64c3a86c480-kube-api-access-q49fx\") pod \"nova-cell1-db-create-fjg6t\" (UID: \"d3954c8e-9c73-4bb3-9db6-e64c3a86c480\") " pod="openstack/nova-cell1-db-create-fjg6t" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.270431 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfphn\" (UniqueName: \"kubernetes.io/projected/d436f2fa-66ed-4acb-970e-8b1116250622-kube-api-access-tfphn\") pod \"nova-cell0-4bc5-account-create-update-d58fg\" (UID: \"d436f2fa-66ed-4acb-970e-8b1116250622\") " pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.270521 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d436f2fa-66ed-4acb-970e-8b1116250622-operator-scripts\") pod \"nova-cell0-4bc5-account-create-update-d58fg\" (UID: \"d436f2fa-66ed-4acb-970e-8b1116250622\") " pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.274754 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fjg6t" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.367529 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-8243-account-create-update-gwqg5"] Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.371716 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8243-account-create-update-gwqg5" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.373786 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d436f2fa-66ed-4acb-970e-8b1116250622-operator-scripts\") pod \"nova-cell0-4bc5-account-create-update-d58fg\" (UID: \"d436f2fa-66ed-4acb-970e-8b1116250622\") " pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.373906 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfphn\" (UniqueName: \"kubernetes.io/projected/d436f2fa-66ed-4acb-970e-8b1116250622-kube-api-access-tfphn\") pod \"nova-cell0-4bc5-account-create-update-d58fg\" (UID: \"d436f2fa-66ed-4acb-970e-8b1116250622\") " pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.374907 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d436f2fa-66ed-4acb-970e-8b1116250622-operator-scripts\") pod \"nova-cell0-4bc5-account-create-update-d58fg\" (UID: \"d436f2fa-66ed-4acb-970e-8b1116250622\") " pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.375727 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.389528 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8243-account-create-update-gwqg5"] Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.407023 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfphn\" (UniqueName: \"kubernetes.io/projected/d436f2fa-66ed-4acb-970e-8b1116250622-kube-api-access-tfphn\") pod \"nova-cell0-4bc5-account-create-update-d58fg\" (UID: \"d436f2fa-66ed-4acb-970e-8b1116250622\") " pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.475872 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd4fca93-e550-4dff-a7d1-5efdd3d121b4-operator-scripts\") pod \"nova-cell1-8243-account-create-update-gwqg5\" (UID: \"bd4fca93-e550-4dff-a7d1-5efdd3d121b4\") " pod="openstack/nova-cell1-8243-account-create-update-gwqg5" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.475943 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmpxs\" (UniqueName: \"kubernetes.io/projected/bd4fca93-e550-4dff-a7d1-5efdd3d121b4-kube-api-access-xmpxs\") pod \"nova-cell1-8243-account-create-update-gwqg5\" (UID: \"bd4fca93-e550-4dff-a7d1-5efdd3d121b4\") " pod="openstack/nova-cell1-8243-account-create-update-gwqg5" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.506076 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.517023 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-28sdq"] Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.578676 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd4fca93-e550-4dff-a7d1-5efdd3d121b4-operator-scripts\") pod \"nova-cell1-8243-account-create-update-gwqg5\" (UID: \"bd4fca93-e550-4dff-a7d1-5efdd3d121b4\") " pod="openstack/nova-cell1-8243-account-create-update-gwqg5" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.578799 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmpxs\" (UniqueName: \"kubernetes.io/projected/bd4fca93-e550-4dff-a7d1-5efdd3d121b4-kube-api-access-xmpxs\") pod \"nova-cell1-8243-account-create-update-gwqg5\" (UID: \"bd4fca93-e550-4dff-a7d1-5efdd3d121b4\") " pod="openstack/nova-cell1-8243-account-create-update-gwqg5" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.580523 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd4fca93-e550-4dff-a7d1-5efdd3d121b4-operator-scripts\") pod \"nova-cell1-8243-account-create-update-gwqg5\" (UID: \"bd4fca93-e550-4dff-a7d1-5efdd3d121b4\") " pod="openstack/nova-cell1-8243-account-create-update-gwqg5" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.612589 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmpxs\" (UniqueName: \"kubernetes.io/projected/bd4fca93-e550-4dff-a7d1-5efdd3d121b4-kube-api-access-xmpxs\") pod \"nova-cell1-8243-account-create-update-gwqg5\" (UID: \"bd4fca93-e550-4dff-a7d1-5efdd3d121b4\") " pod="openstack/nova-cell1-8243-account-create-update-gwqg5" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.711852 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8243-account-create-update-gwqg5" Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.803451 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4fbch"] Nov 27 10:39:38 crc kubenswrapper[4806]: W1127 10:39:38.849722 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod727574b8_48dd_4ca6_882f_e84934bf1240.slice/crio-fbb57c7bd1dbfc6dfb0b9f52ca5c8c13e87cf5968ee541cb514249df6a12d1dd WatchSource:0}: Error finding container fbb57c7bd1dbfc6dfb0b9f52ca5c8c13e87cf5968ee541cb514249df6a12d1dd: Status 404 returned error can't find the container with id fbb57c7bd1dbfc6dfb0b9f52ca5c8c13e87cf5968ee541cb514249df6a12d1dd Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.940674 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7795-account-create-update-wrpdm"] Nov 27 10:39:38 crc kubenswrapper[4806]: I1127 10:39:38.965621 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-4bc5-account-create-update-d58fg"] Nov 27 10:39:39 crc kubenswrapper[4806]: W1127 10:39:39.005476 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd436f2fa_66ed_4acb_970e_8b1116250622.slice/crio-808a38e571703b77e78dc46468073ee6eb672164af66e2fb2517c2960bc7201c WatchSource:0}: Error finding container 808a38e571703b77e78dc46468073ee6eb672164af66e2fb2517c2960bc7201c: Status 404 returned error can't find the container with id 808a38e571703b77e78dc46468073ee6eb672164af66e2fb2517c2960bc7201c Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.030352 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fjg6t"] Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.208475 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8243-account-create-update-gwqg5"] Nov 27 10:39:39 crc kubenswrapper[4806]: W1127 10:39:39.246472 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd4fca93_e550_4dff_a7d1_5efdd3d121b4.slice/crio-52fd6654864afb3c202ac37d5ac58707db5aae29a04d3cbe5d55d8844b42db8f WatchSource:0}: Error finding container 52fd6654864afb3c202ac37d5ac58707db5aae29a04d3cbe5d55d8844b42db8f: Status 404 returned error can't find the container with id 52fd6654864afb3c202ac37d5ac58707db5aae29a04d3cbe5d55d8844b42db8f Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.520815 4806 generic.go:334] "Generic (PLEG): container finished" podID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerID="899dcf3d182a47a7127b251270a2b7d444b62796bbcf8c943b8c4eb9f9afb835" exitCode=0 Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.520912 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4091a1a3-33e0-42ac-935a-a96b43770c62","Type":"ContainerDied","Data":"899dcf3d182a47a7127b251270a2b7d444b62796bbcf8c943b8c4eb9f9afb835"} Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.527421 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4fbch" event={"ID":"727574b8-48dd-4ca6-882f-e84934bf1240","Type":"ContainerStarted","Data":"f0444bce19c77f6a68d3d074f742f9ba2ff1217e4d92b5471c13275c0dc86ae1"} Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.527504 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4fbch" event={"ID":"727574b8-48dd-4ca6-882f-e84934bf1240","Type":"ContainerStarted","Data":"fbb57c7bd1dbfc6dfb0b9f52ca5c8c13e87cf5968ee541cb514249df6a12d1dd"} Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.530108 4806 generic.go:334] "Generic (PLEG): container finished" podID="e7a0c344-9e24-43af-badd-b761a5005c56" containerID="f54b3e0a0fd3f20237c30e8fa9756b990de4f6a8be8489d3a1c2b7d823271841" exitCode=0 Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.530186 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-28sdq" event={"ID":"e7a0c344-9e24-43af-badd-b761a5005c56","Type":"ContainerDied","Data":"f54b3e0a0fd3f20237c30e8fa9756b990de4f6a8be8489d3a1c2b7d823271841"} Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.530216 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-28sdq" event={"ID":"e7a0c344-9e24-43af-badd-b761a5005c56","Type":"ContainerStarted","Data":"e337896827c86db85e4c02ec09d2a8512b26f0006dea8e3c554810c0dc82ef3d"} Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.531432 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8243-account-create-update-gwqg5" event={"ID":"bd4fca93-e550-4dff-a7d1-5efdd3d121b4","Type":"ContainerStarted","Data":"52fd6654864afb3c202ac37d5ac58707db5aae29a04d3cbe5d55d8844b42db8f"} Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.534081 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" event={"ID":"d436f2fa-66ed-4acb-970e-8b1116250622","Type":"ContainerStarted","Data":"529e41881d344faac4322ca1b807ae6ca76229cf4a751d348598a8dcec91c6e2"} Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.534109 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" event={"ID":"d436f2fa-66ed-4acb-970e-8b1116250622","Type":"ContainerStarted","Data":"808a38e571703b77e78dc46468073ee6eb672164af66e2fb2517c2960bc7201c"} Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.535483 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fjg6t" event={"ID":"d3954c8e-9c73-4bb3-9db6-e64c3a86c480","Type":"ContainerStarted","Data":"9717ca758a24a7bd89f9db2e4d21063a1025f77be0cd7c27b15e736c0864d702"} Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.535506 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fjg6t" event={"ID":"d3954c8e-9c73-4bb3-9db6-e64c3a86c480","Type":"ContainerStarted","Data":"2281f08c70dc359f4c024980b3c71a98de0863b7c3c524b2b91dcc160ef718b4"} Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.539254 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7795-account-create-update-wrpdm" event={"ID":"f9041e87-4dd0-442d-a056-59d77e5134d0","Type":"ContainerStarted","Data":"3266f576d79cc8f500fe1c8806bde6016ef45abf6860873c153d950491afa787"} Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.539309 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7795-account-create-update-wrpdm" event={"ID":"f9041e87-4dd0-442d-a056-59d77e5134d0","Type":"ContainerStarted","Data":"d99e6760cb857d08417dab1acb2e007504262e745174f2e2366ece3687c0a8e4"} Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.551391 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-4fbch" podStartSLOduration=2.551366531 podStartE2EDuration="2.551366531s" podCreationTimestamp="2025-11-27 10:39:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:39:39.548957515 +0000 UTC m=+1084.135548289" watchObservedRunningTime="2025-11-27 10:39:39.551366531 +0000 UTC m=+1084.137957295" Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.572538 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" podStartSLOduration=1.572503294 podStartE2EDuration="1.572503294s" podCreationTimestamp="2025-11-27 10:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:39:39.567037736 +0000 UTC m=+1084.153628500" watchObservedRunningTime="2025-11-27 10:39:39.572503294 +0000 UTC m=+1084.159094048" Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.599768 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-fjg6t" podStartSLOduration=2.599747852 podStartE2EDuration="2.599747852s" podCreationTimestamp="2025-11-27 10:39:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:39:39.597198783 +0000 UTC m=+1084.183789547" watchObservedRunningTime="2025-11-27 10:39:39.599747852 +0000 UTC m=+1084.186338626" Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.623284 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-8243-account-create-update-gwqg5" podStartSLOduration=1.623266291 podStartE2EDuration="1.623266291s" podCreationTimestamp="2025-11-27 10:39:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:39:39.61845 +0000 UTC m=+1084.205040764" watchObservedRunningTime="2025-11-27 10:39:39.623266291 +0000 UTC m=+1084.209857065" Nov 27 10:39:39 crc kubenswrapper[4806]: I1127 10:39:39.651250 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-7795-account-create-update-wrpdm" podStartSLOduration=2.651220619 podStartE2EDuration="2.651220619s" podCreationTimestamp="2025-11-27 10:39:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:39:39.645402791 +0000 UTC m=+1084.231993545" watchObservedRunningTime="2025-11-27 10:39:39.651220619 +0000 UTC m=+1084.237811383" Nov 27 10:39:40 crc kubenswrapper[4806]: I1127 10:39:40.551996 4806 generic.go:334] "Generic (PLEG): container finished" podID="727574b8-48dd-4ca6-882f-e84934bf1240" containerID="f0444bce19c77f6a68d3d074f742f9ba2ff1217e4d92b5471c13275c0dc86ae1" exitCode=0 Nov 27 10:39:40 crc kubenswrapper[4806]: I1127 10:39:40.552057 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4fbch" event={"ID":"727574b8-48dd-4ca6-882f-e84934bf1240","Type":"ContainerDied","Data":"f0444bce19c77f6a68d3d074f742f9ba2ff1217e4d92b5471c13275c0dc86ae1"} Nov 27 10:39:40 crc kubenswrapper[4806]: I1127 10:39:40.554722 4806 generic.go:334] "Generic (PLEG): container finished" podID="bd4fca93-e550-4dff-a7d1-5efdd3d121b4" containerID="95cf89f7e44dd54684b95eb8164fa293fcc1e2c211c26814b25a739260a51543" exitCode=0 Nov 27 10:39:40 crc kubenswrapper[4806]: I1127 10:39:40.554816 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8243-account-create-update-gwqg5" event={"ID":"bd4fca93-e550-4dff-a7d1-5efdd3d121b4","Type":"ContainerDied","Data":"95cf89f7e44dd54684b95eb8164fa293fcc1e2c211c26814b25a739260a51543"} Nov 27 10:39:40 crc kubenswrapper[4806]: I1127 10:39:40.557105 4806 generic.go:334] "Generic (PLEG): container finished" podID="d436f2fa-66ed-4acb-970e-8b1116250622" containerID="529e41881d344faac4322ca1b807ae6ca76229cf4a751d348598a8dcec91c6e2" exitCode=0 Nov 27 10:39:40 crc kubenswrapper[4806]: I1127 10:39:40.557204 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" event={"ID":"d436f2fa-66ed-4acb-970e-8b1116250622","Type":"ContainerDied","Data":"529e41881d344faac4322ca1b807ae6ca76229cf4a751d348598a8dcec91c6e2"} Nov 27 10:39:40 crc kubenswrapper[4806]: I1127 10:39:40.559562 4806 generic.go:334] "Generic (PLEG): container finished" podID="d3954c8e-9c73-4bb3-9db6-e64c3a86c480" containerID="9717ca758a24a7bd89f9db2e4d21063a1025f77be0cd7c27b15e736c0864d702" exitCode=0 Nov 27 10:39:40 crc kubenswrapper[4806]: I1127 10:39:40.559592 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fjg6t" event={"ID":"d3954c8e-9c73-4bb3-9db6-e64c3a86c480","Type":"ContainerDied","Data":"9717ca758a24a7bd89f9db2e4d21063a1025f77be0cd7c27b15e736c0864d702"} Nov 27 10:39:40 crc kubenswrapper[4806]: I1127 10:39:40.561820 4806 generic.go:334] "Generic (PLEG): container finished" podID="f9041e87-4dd0-442d-a056-59d77e5134d0" containerID="3266f576d79cc8f500fe1c8806bde6016ef45abf6860873c153d950491afa787" exitCode=0 Nov 27 10:39:40 crc kubenswrapper[4806]: I1127 10:39:40.561894 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7795-account-create-update-wrpdm" event={"ID":"f9041e87-4dd0-442d-a056-59d77e5134d0","Type":"ContainerDied","Data":"3266f576d79cc8f500fe1c8806bde6016ef45abf6860873c153d950491afa787"} Nov 27 10:39:41 crc kubenswrapper[4806]: I1127 10:39:41.026474 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-28sdq" Nov 27 10:39:41 crc kubenswrapper[4806]: I1127 10:39:41.141164 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2n9t\" (UniqueName: \"kubernetes.io/projected/e7a0c344-9e24-43af-badd-b761a5005c56-kube-api-access-z2n9t\") pod \"e7a0c344-9e24-43af-badd-b761a5005c56\" (UID: \"e7a0c344-9e24-43af-badd-b761a5005c56\") " Nov 27 10:39:41 crc kubenswrapper[4806]: I1127 10:39:41.141388 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7a0c344-9e24-43af-badd-b761a5005c56-operator-scripts\") pod \"e7a0c344-9e24-43af-badd-b761a5005c56\" (UID: \"e7a0c344-9e24-43af-badd-b761a5005c56\") " Nov 27 10:39:41 crc kubenswrapper[4806]: I1127 10:39:41.142754 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7a0c344-9e24-43af-badd-b761a5005c56-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e7a0c344-9e24-43af-badd-b761a5005c56" (UID: "e7a0c344-9e24-43af-badd-b761a5005c56"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:39:41 crc kubenswrapper[4806]: I1127 10:39:41.152630 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7a0c344-9e24-43af-badd-b761a5005c56-kube-api-access-z2n9t" (OuterVolumeSpecName: "kube-api-access-z2n9t") pod "e7a0c344-9e24-43af-badd-b761a5005c56" (UID: "e7a0c344-9e24-43af-badd-b761a5005c56"). InnerVolumeSpecName "kube-api-access-z2n9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:39:41 crc kubenswrapper[4806]: I1127 10:39:41.245543 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2n9t\" (UniqueName: \"kubernetes.io/projected/e7a0c344-9e24-43af-badd-b761a5005c56-kube-api-access-z2n9t\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:41 crc kubenswrapper[4806]: I1127 10:39:41.245597 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7a0c344-9e24-43af-badd-b761a5005c56-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:41 crc kubenswrapper[4806]: I1127 10:39:41.572283 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-28sdq" Nov 27 10:39:41 crc kubenswrapper[4806]: I1127 10:39:41.578245 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-28sdq" event={"ID":"e7a0c344-9e24-43af-badd-b761a5005c56","Type":"ContainerDied","Data":"e337896827c86db85e4c02ec09d2a8512b26f0006dea8e3c554810c0dc82ef3d"} Nov 27 10:39:41 crc kubenswrapper[4806]: I1127 10:39:41.578317 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e337896827c86db85e4c02ec09d2a8512b26f0006dea8e3c554810c0dc82ef3d" Nov 27 10:39:41 crc kubenswrapper[4806]: I1127 10:39:41.936416 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fjg6t" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.065104 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3954c8e-9c73-4bb3-9db6-e64c3a86c480-operator-scripts\") pod \"d3954c8e-9c73-4bb3-9db6-e64c3a86c480\" (UID: \"d3954c8e-9c73-4bb3-9db6-e64c3a86c480\") " Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.065169 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q49fx\" (UniqueName: \"kubernetes.io/projected/d3954c8e-9c73-4bb3-9db6-e64c3a86c480-kube-api-access-q49fx\") pod \"d3954c8e-9c73-4bb3-9db6-e64c3a86c480\" (UID: \"d3954c8e-9c73-4bb3-9db6-e64c3a86c480\") " Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.066684 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3954c8e-9c73-4bb3-9db6-e64c3a86c480-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3954c8e-9c73-4bb3-9db6-e64c3a86c480" (UID: "d3954c8e-9c73-4bb3-9db6-e64c3a86c480"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.093029 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3954c8e-9c73-4bb3-9db6-e64c3a86c480-kube-api-access-q49fx" (OuterVolumeSpecName: "kube-api-access-q49fx") pod "d3954c8e-9c73-4bb3-9db6-e64c3a86c480" (UID: "d3954c8e-9c73-4bb3-9db6-e64c3a86c480"). InnerVolumeSpecName "kube-api-access-q49fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.167600 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3954c8e-9c73-4bb3-9db6-e64c3a86c480-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.167633 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q49fx\" (UniqueName: \"kubernetes.io/projected/d3954c8e-9c73-4bb3-9db6-e64c3a86c480-kube-api-access-q49fx\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.382016 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8243-account-create-update-gwqg5" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.389903 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.398888 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7795-account-create-update-wrpdm" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.407154 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4fbch" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.474650 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d436f2fa-66ed-4acb-970e-8b1116250622-operator-scripts\") pod \"d436f2fa-66ed-4acb-970e-8b1116250622\" (UID: \"d436f2fa-66ed-4acb-970e-8b1116250622\") " Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.475062 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfphn\" (UniqueName: \"kubernetes.io/projected/d436f2fa-66ed-4acb-970e-8b1116250622-kube-api-access-tfphn\") pod \"d436f2fa-66ed-4acb-970e-8b1116250622\" (UID: \"d436f2fa-66ed-4acb-970e-8b1116250622\") " Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.475107 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/727574b8-48dd-4ca6-882f-e84934bf1240-operator-scripts\") pod \"727574b8-48dd-4ca6-882f-e84934bf1240\" (UID: \"727574b8-48dd-4ca6-882f-e84934bf1240\") " Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.475141 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7l6h\" (UniqueName: \"kubernetes.io/projected/727574b8-48dd-4ca6-882f-e84934bf1240-kube-api-access-t7l6h\") pod \"727574b8-48dd-4ca6-882f-e84934bf1240\" (UID: \"727574b8-48dd-4ca6-882f-e84934bf1240\") " Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.475176 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d436f2fa-66ed-4acb-970e-8b1116250622-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d436f2fa-66ed-4acb-970e-8b1116250622" (UID: "d436f2fa-66ed-4acb-970e-8b1116250622"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.475210 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd4fca93-e550-4dff-a7d1-5efdd3d121b4-operator-scripts\") pod \"bd4fca93-e550-4dff-a7d1-5efdd3d121b4\" (UID: \"bd4fca93-e550-4dff-a7d1-5efdd3d121b4\") " Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.475388 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmpxs\" (UniqueName: \"kubernetes.io/projected/bd4fca93-e550-4dff-a7d1-5efdd3d121b4-kube-api-access-xmpxs\") pod \"bd4fca93-e550-4dff-a7d1-5efdd3d121b4\" (UID: \"bd4fca93-e550-4dff-a7d1-5efdd3d121b4\") " Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.475413 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wglg\" (UniqueName: \"kubernetes.io/projected/f9041e87-4dd0-442d-a056-59d77e5134d0-kube-api-access-9wglg\") pod \"f9041e87-4dd0-442d-a056-59d77e5134d0\" (UID: \"f9041e87-4dd0-442d-a056-59d77e5134d0\") " Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.475466 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9041e87-4dd0-442d-a056-59d77e5134d0-operator-scripts\") pod \"f9041e87-4dd0-442d-a056-59d77e5134d0\" (UID: \"f9041e87-4dd0-442d-a056-59d77e5134d0\") " Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.475535 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/727574b8-48dd-4ca6-882f-e84934bf1240-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "727574b8-48dd-4ca6-882f-e84934bf1240" (UID: "727574b8-48dd-4ca6-882f-e84934bf1240"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.475805 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d436f2fa-66ed-4acb-970e-8b1116250622-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.475824 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/727574b8-48dd-4ca6-882f-e84934bf1240-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.475868 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd4fca93-e550-4dff-a7d1-5efdd3d121b4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bd4fca93-e550-4dff-a7d1-5efdd3d121b4" (UID: "bd4fca93-e550-4dff-a7d1-5efdd3d121b4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.476801 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9041e87-4dd0-442d-a056-59d77e5134d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f9041e87-4dd0-442d-a056-59d77e5134d0" (UID: "f9041e87-4dd0-442d-a056-59d77e5134d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.480493 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d436f2fa-66ed-4acb-970e-8b1116250622-kube-api-access-tfphn" (OuterVolumeSpecName: "kube-api-access-tfphn") pod "d436f2fa-66ed-4acb-970e-8b1116250622" (UID: "d436f2fa-66ed-4acb-970e-8b1116250622"). InnerVolumeSpecName "kube-api-access-tfphn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.481545 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/727574b8-48dd-4ca6-882f-e84934bf1240-kube-api-access-t7l6h" (OuterVolumeSpecName: "kube-api-access-t7l6h") pod "727574b8-48dd-4ca6-882f-e84934bf1240" (UID: "727574b8-48dd-4ca6-882f-e84934bf1240"). InnerVolumeSpecName "kube-api-access-t7l6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.483041 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd4fca93-e550-4dff-a7d1-5efdd3d121b4-kube-api-access-xmpxs" (OuterVolumeSpecName: "kube-api-access-xmpxs") pod "bd4fca93-e550-4dff-a7d1-5efdd3d121b4" (UID: "bd4fca93-e550-4dff-a7d1-5efdd3d121b4"). InnerVolumeSpecName "kube-api-access-xmpxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.483289 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9041e87-4dd0-442d-a056-59d77e5134d0-kube-api-access-9wglg" (OuterVolumeSpecName: "kube-api-access-9wglg") pod "f9041e87-4dd0-442d-a056-59d77e5134d0" (UID: "f9041e87-4dd0-442d-a056-59d77e5134d0"). InnerVolumeSpecName "kube-api-access-9wglg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.576999 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfphn\" (UniqueName: \"kubernetes.io/projected/d436f2fa-66ed-4acb-970e-8b1116250622-kube-api-access-tfphn\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.577038 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7l6h\" (UniqueName: \"kubernetes.io/projected/727574b8-48dd-4ca6-882f-e84934bf1240-kube-api-access-t7l6h\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.577050 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd4fca93-e550-4dff-a7d1-5efdd3d121b4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.577059 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmpxs\" (UniqueName: \"kubernetes.io/projected/bd4fca93-e550-4dff-a7d1-5efdd3d121b4-kube-api-access-xmpxs\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.577068 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wglg\" (UniqueName: \"kubernetes.io/projected/f9041e87-4dd0-442d-a056-59d77e5134d0-kube-api-access-9wglg\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.577076 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9041e87-4dd0-442d-a056-59d77e5134d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.583974 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" event={"ID":"d436f2fa-66ed-4acb-970e-8b1116250622","Type":"ContainerDied","Data":"808a38e571703b77e78dc46468073ee6eb672164af66e2fb2517c2960bc7201c"} Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.584019 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="808a38e571703b77e78dc46468073ee6eb672164af66e2fb2517c2960bc7201c" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.584026 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4bc5-account-create-update-d58fg" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.585676 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fjg6t" event={"ID":"d3954c8e-9c73-4bb3-9db6-e64c3a86c480","Type":"ContainerDied","Data":"2281f08c70dc359f4c024980b3c71a98de0863b7c3c524b2b91dcc160ef718b4"} Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.585751 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2281f08c70dc359f4c024980b3c71a98de0863b7c3c524b2b91dcc160ef718b4" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.585750 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fjg6t" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.587225 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7795-account-create-update-wrpdm" event={"ID":"f9041e87-4dd0-442d-a056-59d77e5134d0","Type":"ContainerDied","Data":"d99e6760cb857d08417dab1acb2e007504262e745174f2e2366ece3687c0a8e4"} Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.587289 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d99e6760cb857d08417dab1acb2e007504262e745174f2e2366ece3687c0a8e4" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.587260 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7795-account-create-update-wrpdm" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.589053 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4fbch" event={"ID":"727574b8-48dd-4ca6-882f-e84934bf1240","Type":"ContainerDied","Data":"fbb57c7bd1dbfc6dfb0b9f52ca5c8c13e87cf5968ee541cb514249df6a12d1dd"} Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.589073 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4fbch" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.589086 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbb57c7bd1dbfc6dfb0b9f52ca5c8c13e87cf5968ee541cb514249df6a12d1dd" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.591275 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8243-account-create-update-gwqg5" event={"ID":"bd4fca93-e550-4dff-a7d1-5efdd3d121b4","Type":"ContainerDied","Data":"52fd6654864afb3c202ac37d5ac58707db5aae29a04d3cbe5d55d8844b42db8f"} Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.591305 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52fd6654864afb3c202ac37d5ac58707db5aae29a04d3cbe5d55d8844b42db8f" Nov 27 10:39:42 crc kubenswrapper[4806]: I1127 10:39:42.591365 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8243-account-create-update-gwqg5" Nov 27 10:39:44 crc kubenswrapper[4806]: I1127 10:39:44.451049 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:39:44 crc kubenswrapper[4806]: I1127 10:39:44.451409 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:39:44 crc kubenswrapper[4806]: I1127 10:39:44.927369 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.195975 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bj4gg"] Nov 27 10:39:48 crc kubenswrapper[4806]: E1127 10:39:48.196965 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd4fca93-e550-4dff-a7d1-5efdd3d121b4" containerName="mariadb-account-create-update" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.196984 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd4fca93-e550-4dff-a7d1-5efdd3d121b4" containerName="mariadb-account-create-update" Nov 27 10:39:48 crc kubenswrapper[4806]: E1127 10:39:48.196994 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d436f2fa-66ed-4acb-970e-8b1116250622" containerName="mariadb-account-create-update" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.197000 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d436f2fa-66ed-4acb-970e-8b1116250622" containerName="mariadb-account-create-update" Nov 27 10:39:48 crc kubenswrapper[4806]: E1127 10:39:48.197022 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9041e87-4dd0-442d-a056-59d77e5134d0" containerName="mariadb-account-create-update" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.197028 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9041e87-4dd0-442d-a056-59d77e5134d0" containerName="mariadb-account-create-update" Nov 27 10:39:48 crc kubenswrapper[4806]: E1127 10:39:48.197036 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a0c344-9e24-43af-badd-b761a5005c56" containerName="mariadb-database-create" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.197042 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a0c344-9e24-43af-badd-b761a5005c56" containerName="mariadb-database-create" Nov 27 10:39:48 crc kubenswrapper[4806]: E1127 10:39:48.197690 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727574b8-48dd-4ca6-882f-e84934bf1240" containerName="mariadb-database-create" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.197709 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="727574b8-48dd-4ca6-882f-e84934bf1240" containerName="mariadb-database-create" Nov 27 10:39:48 crc kubenswrapper[4806]: E1127 10:39:48.197724 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3954c8e-9c73-4bb3-9db6-e64c3a86c480" containerName="mariadb-database-create" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.197731 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3954c8e-9c73-4bb3-9db6-e64c3a86c480" containerName="mariadb-database-create" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.197964 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d436f2fa-66ed-4acb-970e-8b1116250622" containerName="mariadb-account-create-update" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.197991 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd4fca93-e550-4dff-a7d1-5efdd3d121b4" containerName="mariadb-account-create-update" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.198006 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9041e87-4dd0-442d-a056-59d77e5134d0" containerName="mariadb-account-create-update" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.198018 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7a0c344-9e24-43af-badd-b761a5005c56" containerName="mariadb-database-create" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.198029 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3954c8e-9c73-4bb3-9db6-e64c3a86c480" containerName="mariadb-database-create" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.198039 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="727574b8-48dd-4ca6-882f-e84934bf1240" containerName="mariadb-database-create" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.199254 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.213869 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-79g2x" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.214113 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.214343 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.224474 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bj4gg"] Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.286443 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bj4gg\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.286507 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4kdh\" (UniqueName: \"kubernetes.io/projected/d7701f56-4e55-43dc-a526-3775fff2526c-kube-api-access-f4kdh\") pod \"nova-cell0-conductor-db-sync-bj4gg\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.286661 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-scripts\") pod \"nova-cell0-conductor-db-sync-bj4gg\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.286741 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-config-data\") pod \"nova-cell0-conductor-db-sync-bj4gg\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.389094 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-scripts\") pod \"nova-cell0-conductor-db-sync-bj4gg\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.389145 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-config-data\") pod \"nova-cell0-conductor-db-sync-bj4gg\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.389286 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bj4gg\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.389307 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4kdh\" (UniqueName: \"kubernetes.io/projected/d7701f56-4e55-43dc-a526-3775fff2526c-kube-api-access-f4kdh\") pod \"nova-cell0-conductor-db-sync-bj4gg\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.398136 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bj4gg\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.399898 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-scripts\") pod \"nova-cell0-conductor-db-sync-bj4gg\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.400439 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-config-data\") pod \"nova-cell0-conductor-db-sync-bj4gg\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.418796 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4kdh\" (UniqueName: \"kubernetes.io/projected/d7701f56-4e55-43dc-a526-3775fff2526c-kube-api-access-f4kdh\") pod \"nova-cell0-conductor-db-sync-bj4gg\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:48 crc kubenswrapper[4806]: I1127 10:39:48.542433 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:39:49 crc kubenswrapper[4806]: I1127 10:39:49.020740 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bj4gg"] Nov 27 10:39:49 crc kubenswrapper[4806]: W1127 10:39:49.030420 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7701f56_4e55_43dc_a526_3775fff2526c.slice/crio-5fc3b2ae71d72f81ba16d37339fb6407ee14bb85353ce1b0a32cf7722b725bdc WatchSource:0}: Error finding container 5fc3b2ae71d72f81ba16d37339fb6407ee14bb85353ce1b0a32cf7722b725bdc: Status 404 returned error can't find the container with id 5fc3b2ae71d72f81ba16d37339fb6407ee14bb85353ce1b0a32cf7722b725bdc Nov 27 10:39:49 crc kubenswrapper[4806]: I1127 10:39:49.653601 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bj4gg" event={"ID":"d7701f56-4e55-43dc-a526-3775fff2526c","Type":"ContainerStarted","Data":"5fc3b2ae71d72f81ba16d37339fb6407ee14bb85353ce1b0a32cf7722b725bdc"} Nov 27 10:39:56 crc kubenswrapper[4806]: I1127 10:39:56.734721 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bj4gg" event={"ID":"d7701f56-4e55-43dc-a526-3775fff2526c","Type":"ContainerStarted","Data":"af60cd81fcfdc97a1eac9448be597d2b6886c711e903e6e66d33faec24bd354d"} Nov 27 10:39:56 crc kubenswrapper[4806]: I1127 10:39:56.766937 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-bj4gg" podStartSLOduration=1.643290114 podStartE2EDuration="8.766914521s" podCreationTimestamp="2025-11-27 10:39:48 +0000 UTC" firstStartedPulling="2025-11-27 10:39:49.0335983 +0000 UTC m=+1093.620189064" lastFinishedPulling="2025-11-27 10:39:56.157222707 +0000 UTC m=+1100.743813471" observedRunningTime="2025-11-27 10:39:56.763309723 +0000 UTC m=+1101.349900497" watchObservedRunningTime="2025-11-27 10:39:56.766914521 +0000 UTC m=+1101.353505285" Nov 27 10:40:01 crc kubenswrapper[4806]: I1127 10:40:01.795442 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 27 10:40:06 crc kubenswrapper[4806]: I1127 10:40:06.867708 4806 generic.go:334] "Generic (PLEG): container finished" podID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerID="4d5346db599f0475d4af9dc97a5f424b6acd4a7fddbb47f122d99bc78f6a2e31" exitCode=137 Nov 27 10:40:06 crc kubenswrapper[4806]: I1127 10:40:06.867861 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4091a1a3-33e0-42ac-935a-a96b43770c62","Type":"ContainerDied","Data":"4d5346db599f0475d4af9dc97a5f424b6acd4a7fddbb47f122d99bc78f6a2e31"} Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.394342 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.574646 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4091a1a3-33e0-42ac-935a-a96b43770c62-log-httpd\") pod \"4091a1a3-33e0-42ac-935a-a96b43770c62\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.574837 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-config-data\") pod \"4091a1a3-33e0-42ac-935a-a96b43770c62\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.575045 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-combined-ca-bundle\") pod \"4091a1a3-33e0-42ac-935a-a96b43770c62\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.575181 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb5x8\" (UniqueName: \"kubernetes.io/projected/4091a1a3-33e0-42ac-935a-a96b43770c62-kube-api-access-xb5x8\") pod \"4091a1a3-33e0-42ac-935a-a96b43770c62\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.575275 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4091a1a3-33e0-42ac-935a-a96b43770c62-run-httpd\") pod \"4091a1a3-33e0-42ac-935a-a96b43770c62\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.575334 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-sg-core-conf-yaml\") pod \"4091a1a3-33e0-42ac-935a-a96b43770c62\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.575371 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-scripts\") pod \"4091a1a3-33e0-42ac-935a-a96b43770c62\" (UID: \"4091a1a3-33e0-42ac-935a-a96b43770c62\") " Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.575458 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4091a1a3-33e0-42ac-935a-a96b43770c62-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4091a1a3-33e0-42ac-935a-a96b43770c62" (UID: "4091a1a3-33e0-42ac-935a-a96b43770c62"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.576139 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4091a1a3-33e0-42ac-935a-a96b43770c62-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4091a1a3-33e0-42ac-935a-a96b43770c62" (UID: "4091a1a3-33e0-42ac-935a-a96b43770c62"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.576566 4806 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4091a1a3-33e0-42ac-935a-a96b43770c62-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.576600 4806 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4091a1a3-33e0-42ac-935a-a96b43770c62-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.582487 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4091a1a3-33e0-42ac-935a-a96b43770c62-kube-api-access-xb5x8" (OuterVolumeSpecName: "kube-api-access-xb5x8") pod "4091a1a3-33e0-42ac-935a-a96b43770c62" (UID: "4091a1a3-33e0-42ac-935a-a96b43770c62"). InnerVolumeSpecName "kube-api-access-xb5x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.583328 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-scripts" (OuterVolumeSpecName: "scripts") pod "4091a1a3-33e0-42ac-935a-a96b43770c62" (UID: "4091a1a3-33e0-42ac-935a-a96b43770c62"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.612175 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4091a1a3-33e0-42ac-935a-a96b43770c62" (UID: "4091a1a3-33e0-42ac-935a-a96b43770c62"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.659453 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4091a1a3-33e0-42ac-935a-a96b43770c62" (UID: "4091a1a3-33e0-42ac-935a-a96b43770c62"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.679081 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb5x8\" (UniqueName: \"kubernetes.io/projected/4091a1a3-33e0-42ac-935a-a96b43770c62-kube-api-access-xb5x8\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.679399 4806 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.679467 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.679523 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.699873 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-config-data" (OuterVolumeSpecName: "config-data") pod "4091a1a3-33e0-42ac-935a-a96b43770c62" (UID: "4091a1a3-33e0-42ac-935a-a96b43770c62"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.782092 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4091a1a3-33e0-42ac-935a-a96b43770c62-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.883034 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4091a1a3-33e0-42ac-935a-a96b43770c62","Type":"ContainerDied","Data":"e75e6688f07e54c2a84da523caaa690dbbd9e1a7424fa935677f27b1115faf57"} Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.883100 4806 scope.go:117] "RemoveContainer" containerID="4d5346db599f0475d4af9dc97a5f424b6acd4a7fddbb47f122d99bc78f6a2e31" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.884694 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.924218 4806 scope.go:117] "RemoveContainer" containerID="e9da1debb9c953660dd0afee85de3504760da50ecc3dbaeee4041c629f8e3c47" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.935564 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.943883 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.961679 4806 scope.go:117] "RemoveContainer" containerID="08c19d395ef8d5c9a9fecf6d0ef362b2afdc4edac7489f8230730d4c22fbb5b9" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.972194 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:40:07 crc kubenswrapper[4806]: E1127 10:40:07.974443 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="ceilometer-central-agent" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.974601 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="ceilometer-central-agent" Nov 27 10:40:07 crc kubenswrapper[4806]: E1127 10:40:07.974698 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="sg-core" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.974764 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="sg-core" Nov 27 10:40:07 crc kubenswrapper[4806]: E1127 10:40:07.974816 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="proxy-httpd" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.974862 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="proxy-httpd" Nov 27 10:40:07 crc kubenswrapper[4806]: E1127 10:40:07.974924 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="ceilometer-notification-agent" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.974978 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="ceilometer-notification-agent" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.975207 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="ceilometer-notification-agent" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.975292 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="ceilometer-central-agent" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.975349 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="sg-core" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.975403 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" containerName="proxy-httpd" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.977143 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:40:07 crc kubenswrapper[4806]: I1127 10:40:07.980429 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.001731 4806 scope.go:117] "RemoveContainer" containerID="899dcf3d182a47a7127b251270a2b7d444b62796bbcf8c943b8c4eb9f9afb835" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.002071 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.021928 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.091561 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.091871 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c820267-b03d-4918-b7fd-aa30d9205034-run-httpd\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.092032 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c820267-b03d-4918-b7fd-aa30d9205034-log-httpd\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.092103 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-scripts\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.092368 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvsv4\" (UniqueName: \"kubernetes.io/projected/3c820267-b03d-4918-b7fd-aa30d9205034-kube-api-access-xvsv4\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.092573 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-config-data\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.092602 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.133757 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4091a1a3-33e0-42ac-935a-a96b43770c62" path="/var/lib/kubelet/pods/4091a1a3-33e0-42ac-935a-a96b43770c62/volumes" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.194884 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.194941 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c820267-b03d-4918-b7fd-aa30d9205034-run-httpd\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.194990 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c820267-b03d-4918-b7fd-aa30d9205034-log-httpd\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.195030 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-scripts\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.195075 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvsv4\" (UniqueName: \"kubernetes.io/projected/3c820267-b03d-4918-b7fd-aa30d9205034-kube-api-access-xvsv4\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.195152 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-config-data\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.195179 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.196927 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c820267-b03d-4918-b7fd-aa30d9205034-log-httpd\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.197159 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c820267-b03d-4918-b7fd-aa30d9205034-run-httpd\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.203680 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-config-data\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.206355 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.209454 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.211129 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-scripts\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.214988 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvsv4\" (UniqueName: \"kubernetes.io/projected/3c820267-b03d-4918-b7fd-aa30d9205034-kube-api-access-xvsv4\") pod \"ceilometer-0\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.306562 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.779046 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:40:08 crc kubenswrapper[4806]: I1127 10:40:08.897289 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c820267-b03d-4918-b7fd-aa30d9205034","Type":"ContainerStarted","Data":"b019ca168fcb15082699af8505faab7b40002d6538a63c94c573df8a23a9ebc7"} Nov 27 10:40:09 crc kubenswrapper[4806]: I1127 10:40:09.906490 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c820267-b03d-4918-b7fd-aa30d9205034","Type":"ContainerStarted","Data":"0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164"} Nov 27 10:40:10 crc kubenswrapper[4806]: I1127 10:40:10.920428 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c820267-b03d-4918-b7fd-aa30d9205034","Type":"ContainerStarted","Data":"e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97"} Nov 27 10:40:10 crc kubenswrapper[4806]: I1127 10:40:10.922536 4806 generic.go:334] "Generic (PLEG): container finished" podID="d7701f56-4e55-43dc-a526-3775fff2526c" containerID="af60cd81fcfdc97a1eac9448be597d2b6886c711e903e6e66d33faec24bd354d" exitCode=0 Nov 27 10:40:10 crc kubenswrapper[4806]: I1127 10:40:10.922578 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bj4gg" event={"ID":"d7701f56-4e55-43dc-a526-3775fff2526c","Type":"ContainerDied","Data":"af60cd81fcfdc97a1eac9448be597d2b6886c711e903e6e66d33faec24bd354d"} Nov 27 10:40:11 crc kubenswrapper[4806]: I1127 10:40:11.936469 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c820267-b03d-4918-b7fd-aa30d9205034","Type":"ContainerStarted","Data":"b5f0194541e5bb8ceccb08213ca8d9a563c56563f016c629db20ff57047f9dda"} Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.370125 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.474760 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-config-data\") pod \"d7701f56-4e55-43dc-a526-3775fff2526c\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.474853 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-combined-ca-bundle\") pod \"d7701f56-4e55-43dc-a526-3775fff2526c\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.474955 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4kdh\" (UniqueName: \"kubernetes.io/projected/d7701f56-4e55-43dc-a526-3775fff2526c-kube-api-access-f4kdh\") pod \"d7701f56-4e55-43dc-a526-3775fff2526c\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.475084 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-scripts\") pod \"d7701f56-4e55-43dc-a526-3775fff2526c\" (UID: \"d7701f56-4e55-43dc-a526-3775fff2526c\") " Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.479036 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-scripts" (OuterVolumeSpecName: "scripts") pod "d7701f56-4e55-43dc-a526-3775fff2526c" (UID: "d7701f56-4e55-43dc-a526-3775fff2526c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.481359 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7701f56-4e55-43dc-a526-3775fff2526c-kube-api-access-f4kdh" (OuterVolumeSpecName: "kube-api-access-f4kdh") pod "d7701f56-4e55-43dc-a526-3775fff2526c" (UID: "d7701f56-4e55-43dc-a526-3775fff2526c"). InnerVolumeSpecName "kube-api-access-f4kdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.505599 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7701f56-4e55-43dc-a526-3775fff2526c" (UID: "d7701f56-4e55-43dc-a526-3775fff2526c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.507151 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-config-data" (OuterVolumeSpecName: "config-data") pod "d7701f56-4e55-43dc-a526-3775fff2526c" (UID: "d7701f56-4e55-43dc-a526-3775fff2526c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.577690 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4kdh\" (UniqueName: \"kubernetes.io/projected/d7701f56-4e55-43dc-a526-3775fff2526c-kube-api-access-f4kdh\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.577727 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.577738 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.577751 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7701f56-4e55-43dc-a526-3775fff2526c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.960763 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bj4gg" event={"ID":"d7701f56-4e55-43dc-a526-3775fff2526c","Type":"ContainerDied","Data":"5fc3b2ae71d72f81ba16d37339fb6407ee14bb85353ce1b0a32cf7722b725bdc"} Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.961038 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fc3b2ae71d72f81ba16d37339fb6407ee14bb85353ce1b0a32cf7722b725bdc" Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.961109 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bj4gg" Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.973861 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c820267-b03d-4918-b7fd-aa30d9205034","Type":"ContainerStarted","Data":"d3f52d60ec4e75a0974a6d083f8492ad4a50d8e1228ff9e15a20355a9c5bf5ef"} Nov 27 10:40:12 crc kubenswrapper[4806]: I1127 10:40:12.974634 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.001661 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.566653274 podStartE2EDuration="6.001636373s" podCreationTimestamp="2025-11-27 10:40:07 +0000 UTC" firstStartedPulling="2025-11-27 10:40:08.801513034 +0000 UTC m=+1113.388103798" lastFinishedPulling="2025-11-27 10:40:12.236496133 +0000 UTC m=+1116.823086897" observedRunningTime="2025-11-27 10:40:12.997746527 +0000 UTC m=+1117.584337301" watchObservedRunningTime="2025-11-27 10:40:13.001636373 +0000 UTC m=+1117.588227137" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.064824 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 27 10:40:13 crc kubenswrapper[4806]: E1127 10:40:13.065952 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7701f56-4e55-43dc-a526-3775fff2526c" containerName="nova-cell0-conductor-db-sync" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.065988 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7701f56-4e55-43dc-a526-3775fff2526c" containerName="nova-cell0-conductor-db-sync" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.067704 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7701f56-4e55-43dc-a526-3775fff2526c" containerName="nova-cell0-conductor-db-sync" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.074714 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.080916 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-79g2x" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.081288 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.145709 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.215469 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/570b138c-102e-4d40-a2df-a87abf1626b7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"570b138c-102e-4d40-a2df-a87abf1626b7\") " pod="openstack/nova-cell0-conductor-0" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.215525 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/570b138c-102e-4d40-a2df-a87abf1626b7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"570b138c-102e-4d40-a2df-a87abf1626b7\") " pod="openstack/nova-cell0-conductor-0" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.215567 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqgmb\" (UniqueName: \"kubernetes.io/projected/570b138c-102e-4d40-a2df-a87abf1626b7-kube-api-access-rqgmb\") pod \"nova-cell0-conductor-0\" (UID: \"570b138c-102e-4d40-a2df-a87abf1626b7\") " pod="openstack/nova-cell0-conductor-0" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.317948 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/570b138c-102e-4d40-a2df-a87abf1626b7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"570b138c-102e-4d40-a2df-a87abf1626b7\") " pod="openstack/nova-cell0-conductor-0" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.318013 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/570b138c-102e-4d40-a2df-a87abf1626b7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"570b138c-102e-4d40-a2df-a87abf1626b7\") " pod="openstack/nova-cell0-conductor-0" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.318033 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqgmb\" (UniqueName: \"kubernetes.io/projected/570b138c-102e-4d40-a2df-a87abf1626b7-kube-api-access-rqgmb\") pod \"nova-cell0-conductor-0\" (UID: \"570b138c-102e-4d40-a2df-a87abf1626b7\") " pod="openstack/nova-cell0-conductor-0" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.321962 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/570b138c-102e-4d40-a2df-a87abf1626b7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"570b138c-102e-4d40-a2df-a87abf1626b7\") " pod="openstack/nova-cell0-conductor-0" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.322024 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/570b138c-102e-4d40-a2df-a87abf1626b7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"570b138c-102e-4d40-a2df-a87abf1626b7\") " pod="openstack/nova-cell0-conductor-0" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.332963 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqgmb\" (UniqueName: \"kubernetes.io/projected/570b138c-102e-4d40-a2df-a87abf1626b7-kube-api-access-rqgmb\") pod \"nova-cell0-conductor-0\" (UID: \"570b138c-102e-4d40-a2df-a87abf1626b7\") " pod="openstack/nova-cell0-conductor-0" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.411298 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.856824 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 27 10:40:13 crc kubenswrapper[4806]: I1127 10:40:13.997470 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"570b138c-102e-4d40-a2df-a87abf1626b7","Type":"ContainerStarted","Data":"a54b224ff1c6f6a7ec7efc6fcb56255d1fe3258ab067c2b8d8025d66503d3cfd"} Nov 27 10:40:14 crc kubenswrapper[4806]: I1127 10:40:14.450445 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:40:14 crc kubenswrapper[4806]: I1127 10:40:14.450512 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:40:14 crc kubenswrapper[4806]: I1127 10:40:14.450566 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:40:14 crc kubenswrapper[4806]: I1127 10:40:14.451358 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5908b1682df3b6a3dc72ea1b0d3756efb806c890ef007bbc5b48b3a697274cc4"} pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 10:40:14 crc kubenswrapper[4806]: I1127 10:40:14.451433 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" containerID="cri-o://5908b1682df3b6a3dc72ea1b0d3756efb806c890ef007bbc5b48b3a697274cc4" gracePeriod=600 Nov 27 10:40:15 crc kubenswrapper[4806]: I1127 10:40:15.011251 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"570b138c-102e-4d40-a2df-a87abf1626b7","Type":"ContainerStarted","Data":"d7e0366b30f8170853a67c6007862031b45d3357b7fab984bc39f1447639bb2d"} Nov 27 10:40:15 crc kubenswrapper[4806]: I1127 10:40:15.012127 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 27 10:40:15 crc kubenswrapper[4806]: I1127 10:40:15.018250 4806 generic.go:334] "Generic (PLEG): container finished" podID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerID="5908b1682df3b6a3dc72ea1b0d3756efb806c890ef007bbc5b48b3a697274cc4" exitCode=0 Nov 27 10:40:15 crc kubenswrapper[4806]: I1127 10:40:15.018315 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerDied","Data":"5908b1682df3b6a3dc72ea1b0d3756efb806c890ef007bbc5b48b3a697274cc4"} Nov 27 10:40:15 crc kubenswrapper[4806]: I1127 10:40:15.018367 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"e9030989914737d800499533c98818930f7560ca5f3c521acf106104ae9ed373"} Nov 27 10:40:15 crc kubenswrapper[4806]: I1127 10:40:15.018388 4806 scope.go:117] "RemoveContainer" containerID="dc655ac750298c594d4faac6cef9782fc2d5a20661ea69277b9fafe78a2e81f0" Nov 27 10:40:15 crc kubenswrapper[4806]: I1127 10:40:15.032791 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.032761693 podStartE2EDuration="2.032761693s" podCreationTimestamp="2025-11-27 10:40:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:40:15.031373385 +0000 UTC m=+1119.617964149" watchObservedRunningTime="2025-11-27 10:40:15.032761693 +0000 UTC m=+1119.619352457" Nov 27 10:40:23 crc kubenswrapper[4806]: I1127 10:40:23.438318 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 27 10:40:23 crc kubenswrapper[4806]: I1127 10:40:23.868891 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-zhv9q"] Nov 27 10:40:23 crc kubenswrapper[4806]: I1127 10:40:23.870396 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:23 crc kubenswrapper[4806]: I1127 10:40:23.876696 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 27 10:40:23 crc kubenswrapper[4806]: I1127 10:40:23.876894 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 27 10:40:23 crc kubenswrapper[4806]: I1127 10:40:23.887208 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zhv9q"] Nov 27 10:40:23 crc kubenswrapper[4806]: I1127 10:40:23.947463 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zhv9q\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:23 crc kubenswrapper[4806]: I1127 10:40:23.947563 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-scripts\") pod \"nova-cell0-cell-mapping-zhv9q\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:23 crc kubenswrapper[4806]: I1127 10:40:23.947596 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66xcd\" (UniqueName: \"kubernetes.io/projected/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-kube-api-access-66xcd\") pod \"nova-cell0-cell-mapping-zhv9q\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:23 crc kubenswrapper[4806]: I1127 10:40:23.947672 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-config-data\") pod \"nova-cell0-cell-mapping-zhv9q\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.048443 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.049368 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zhv9q\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.049433 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-scripts\") pod \"nova-cell0-cell-mapping-zhv9q\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.049454 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66xcd\" (UniqueName: \"kubernetes.io/projected/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-kube-api-access-66xcd\") pod \"nova-cell0-cell-mapping-zhv9q\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.049501 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-config-data\") pod \"nova-cell0-cell-mapping-zhv9q\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.057112 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.065532 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.066474 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-scripts\") pod \"nova-cell0-cell-mapping-zhv9q\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.067392 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-config-data\") pod \"nova-cell0-cell-mapping-zhv9q\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.073512 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zhv9q\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.105978 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.148772 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66xcd\" (UniqueName: \"kubernetes.io/projected/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-kube-api-access-66xcd\") pod \"nova-cell0-cell-mapping-zhv9q\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.150688 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nnv2\" (UniqueName: \"kubernetes.io/projected/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-kube-api-access-9nnv2\") pod \"nova-api-0\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.150885 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-logs\") pod \"nova-api-0\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.151039 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-config-data\") pod \"nova-api-0\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.151126 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.176272 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.177472 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.204539 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.224991 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.231280 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.252479 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-logs\") pod \"nova-api-0\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.253300 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-logs\") pod \"nova-api-0\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.253380 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-config-data\") pod \"nova-api-0\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.253440 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.253793 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nnv2\" (UniqueName: \"kubernetes.io/projected/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-kube-api-access-9nnv2\") pod \"nova-api-0\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.265779 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-config-data\") pod \"nova-api-0\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.299122 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.299363 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nnv2\" (UniqueName: \"kubernetes.io/projected/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-kube-api-access-9nnv2\") pod \"nova-api-0\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.302825 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.304002 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.313756 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.359955 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.385515 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.407902 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1363ac1b-7673-4792-82d0-cbccdbd881cb-config-data\") pod \"nova-scheduler-0\" (UID: \"1363ac1b-7673-4792-82d0-cbccdbd881cb\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.409101 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794a9a51-b1a0-49c9-8308-d71995e9e23a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"794a9a51-b1a0-49c9-8308-d71995e9e23a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.409459 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794a9a51-b1a0-49c9-8308-d71995e9e23a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"794a9a51-b1a0-49c9-8308-d71995e9e23a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.409785 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5xjq\" (UniqueName: \"kubernetes.io/projected/794a9a51-b1a0-49c9-8308-d71995e9e23a-kube-api-access-q5xjq\") pod \"nova-cell1-novncproxy-0\" (UID: \"794a9a51-b1a0-49c9-8308-d71995e9e23a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.410266 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wxrm\" (UniqueName: \"kubernetes.io/projected/1363ac1b-7673-4792-82d0-cbccdbd881cb-kube-api-access-5wxrm\") pod \"nova-scheduler-0\" (UID: \"1363ac1b-7673-4792-82d0-cbccdbd881cb\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.410598 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1363ac1b-7673-4792-82d0-cbccdbd881cb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1363ac1b-7673-4792-82d0-cbccdbd881cb\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.516717 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1363ac1b-7673-4792-82d0-cbccdbd881cb-config-data\") pod \"nova-scheduler-0\" (UID: \"1363ac1b-7673-4792-82d0-cbccdbd881cb\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.516952 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.516841 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794a9a51-b1a0-49c9-8308-d71995e9e23a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"794a9a51-b1a0-49c9-8308-d71995e9e23a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.521247 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794a9a51-b1a0-49c9-8308-d71995e9e23a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"794a9a51-b1a0-49c9-8308-d71995e9e23a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.521386 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5xjq\" (UniqueName: \"kubernetes.io/projected/794a9a51-b1a0-49c9-8308-d71995e9e23a-kube-api-access-q5xjq\") pod \"nova-cell1-novncproxy-0\" (UID: \"794a9a51-b1a0-49c9-8308-d71995e9e23a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.532912 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wxrm\" (UniqueName: \"kubernetes.io/projected/1363ac1b-7673-4792-82d0-cbccdbd881cb-kube-api-access-5wxrm\") pod \"nova-scheduler-0\" (UID: \"1363ac1b-7673-4792-82d0-cbccdbd881cb\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.532959 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.533068 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1363ac1b-7673-4792-82d0-cbccdbd881cb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1363ac1b-7673-4792-82d0-cbccdbd881cb\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.541807 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.550338 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794a9a51-b1a0-49c9-8308-d71995e9e23a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"794a9a51-b1a0-49c9-8308-d71995e9e23a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.564650 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1363ac1b-7673-4792-82d0-cbccdbd881cb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1363ac1b-7673-4792-82d0-cbccdbd881cb\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.565099 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1363ac1b-7673-4792-82d0-cbccdbd881cb-config-data\") pod \"nova-scheduler-0\" (UID: \"1363ac1b-7673-4792-82d0-cbccdbd881cb\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.579540 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wxrm\" (UniqueName: \"kubernetes.io/projected/1363ac1b-7673-4792-82d0-cbccdbd881cb-kube-api-access-5wxrm\") pod \"nova-scheduler-0\" (UID: \"1363ac1b-7673-4792-82d0-cbccdbd881cb\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.579617 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.585098 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-tdl4c"] Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.586783 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.590934 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794a9a51-b1a0-49c9-8308-d71995e9e23a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"794a9a51-b1a0-49c9-8308-d71995e9e23a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.602403 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5xjq\" (UniqueName: \"kubernetes.io/projected/794a9a51-b1a0-49c9-8308-d71995e9e23a-kube-api-access-q5xjq\") pod \"nova-cell1-novncproxy-0\" (UID: \"794a9a51-b1a0-49c9-8308-d71995e9e23a\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.637699 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-config\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.637743 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.637771 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k5vf\" (UniqueName: \"kubernetes.io/projected/d5f6cb41-571f-44d9-b269-33d80225af04-kube-api-access-6k5vf\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.637803 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2147d816-df60-4343-b25d-901376f1f90f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.637845 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2147d816-df60-4343-b25d-901376f1f90f-config-data\") pod \"nova-metadata-0\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.637915 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.637951 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.637972 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2147d816-df60-4343-b25d-901376f1f90f-logs\") pod \"nova-metadata-0\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.638000 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h7sn\" (UniqueName: \"kubernetes.io/projected/2147d816-df60-4343-b25d-901376f1f90f-kube-api-access-5h7sn\") pod \"nova-metadata-0\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.707051 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-tdl4c"] Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.712117 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.739198 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.739299 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k5vf\" (UniqueName: \"kubernetes.io/projected/d5f6cb41-571f-44d9-b269-33d80225af04-kube-api-access-6k5vf\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.739351 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2147d816-df60-4343-b25d-901376f1f90f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.739402 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2147d816-df60-4343-b25d-901376f1f90f-config-data\") pod \"nova-metadata-0\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.739488 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.739543 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.739572 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2147d816-df60-4343-b25d-901376f1f90f-logs\") pod \"nova-metadata-0\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.739616 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h7sn\" (UniqueName: \"kubernetes.io/projected/2147d816-df60-4343-b25d-901376f1f90f-kube-api-access-5h7sn\") pod \"nova-metadata-0\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.739657 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-config\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.741725 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.744208 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-config\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.744629 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2147d816-df60-4343-b25d-901376f1f90f-logs\") pod \"nova-metadata-0\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.746839 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.754872 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.755384 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.756952 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2147d816-df60-4343-b25d-901376f1f90f-config-data\") pod \"nova-metadata-0\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.771486 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2147d816-df60-4343-b25d-901376f1f90f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.775021 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h7sn\" (UniqueName: \"kubernetes.io/projected/2147d816-df60-4343-b25d-901376f1f90f-kube-api-access-5h7sn\") pod \"nova-metadata-0\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.775942 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k5vf\" (UniqueName: \"kubernetes.io/projected/d5f6cb41-571f-44d9-b269-33d80225af04-kube-api-access-6k5vf\") pod \"dnsmasq-dns-8b8cf6657-tdl4c\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.886023 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 10:40:24 crc kubenswrapper[4806]: I1127 10:40:24.925647 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.011703 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:40:25 crc kubenswrapper[4806]: W1127 10:40:25.024386 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ccbc1ed_f210_4db1_a8a4_aa022a9913dd.slice/crio-9b398ac377e4aee8cf9478a41e03a8484f9f026c3d14977902ce08a681f32f3e WatchSource:0}: Error finding container 9b398ac377e4aee8cf9478a41e03a8484f9f026c3d14977902ce08a681f32f3e: Status 404 returned error can't find the container with id 9b398ac377e4aee8cf9478a41e03a8484f9f026c3d14977902ce08a681f32f3e Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.128639 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zhv9q"] Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.142359 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd","Type":"ContainerStarted","Data":"9b398ac377e4aee8cf9478a41e03a8484f9f026c3d14977902ce08a681f32f3e"} Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.462948 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.517896 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 10:40:25 crc kubenswrapper[4806]: W1127 10:40:25.527279 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod794a9a51_b1a0_49c9_8308_d71995e9e23a.slice/crio-f752a3c5dc361d0c443c9583f70f59fe8ac3cdb8cfeb831ca60ee16d2aa89e3e WatchSource:0}: Error finding container f752a3c5dc361d0c443c9583f70f59fe8ac3cdb8cfeb831ca60ee16d2aa89e3e: Status 404 returned error can't find the container with id f752a3c5dc361d0c443c9583f70f59fe8ac3cdb8cfeb831ca60ee16d2aa89e3e Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.571483 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6j7m6"] Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.572693 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.576433 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.581761 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.586212 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6j7m6"] Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.665664 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-tdl4c"] Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.665879 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-scripts\") pod \"nova-cell1-conductor-db-sync-6j7m6\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.665944 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6j7m6\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.666054 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rfld\" (UniqueName: \"kubernetes.io/projected/2b0f7898-ee70-46b9-a405-54419ec98b47-kube-api-access-9rfld\") pod \"nova-cell1-conductor-db-sync-6j7m6\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.666154 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-config-data\") pod \"nova-cell1-conductor-db-sync-6j7m6\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.703318 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.780214 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rfld\" (UniqueName: \"kubernetes.io/projected/2b0f7898-ee70-46b9-a405-54419ec98b47-kube-api-access-9rfld\") pod \"nova-cell1-conductor-db-sync-6j7m6\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.780504 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-config-data\") pod \"nova-cell1-conductor-db-sync-6j7m6\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.783440 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-scripts\") pod \"nova-cell1-conductor-db-sync-6j7m6\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.783824 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6j7m6\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.792483 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-config-data\") pod \"nova-cell1-conductor-db-sync-6j7m6\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.795748 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6j7m6\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.797757 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-scripts\") pod \"nova-cell1-conductor-db-sync-6j7m6\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.807842 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rfld\" (UniqueName: \"kubernetes.io/projected/2b0f7898-ee70-46b9-a405-54419ec98b47-kube-api-access-9rfld\") pod \"nova-cell1-conductor-db-sync-6j7m6\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:25 crc kubenswrapper[4806]: I1127 10:40:25.890015 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:26 crc kubenswrapper[4806]: I1127 10:40:26.205734 4806 generic.go:334] "Generic (PLEG): container finished" podID="d5f6cb41-571f-44d9-b269-33d80225af04" containerID="ac0b3b755c9861d50dee1dcfb9431823ffb45a93a2d9881ab7fc7532d4cb3463" exitCode=0 Nov 27 10:40:26 crc kubenswrapper[4806]: I1127 10:40:26.205792 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" event={"ID":"d5f6cb41-571f-44d9-b269-33d80225af04","Type":"ContainerDied","Data":"ac0b3b755c9861d50dee1dcfb9431823ffb45a93a2d9881ab7fc7532d4cb3463"} Nov 27 10:40:26 crc kubenswrapper[4806]: I1127 10:40:26.205816 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" event={"ID":"d5f6cb41-571f-44d9-b269-33d80225af04","Type":"ContainerStarted","Data":"27c067be34ba9a361855eca26c9478d894489ac7b0fc74fa635b05d0e0e04a8c"} Nov 27 10:40:26 crc kubenswrapper[4806]: I1127 10:40:26.208974 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1363ac1b-7673-4792-82d0-cbccdbd881cb","Type":"ContainerStarted","Data":"4cd538a9ca86ada617893f51760d29e64903bf77826030ffe83f31452f8fb114"} Nov 27 10:40:26 crc kubenswrapper[4806]: I1127 10:40:26.222693 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2147d816-df60-4343-b25d-901376f1f90f","Type":"ContainerStarted","Data":"59ceed370c930105930798f821e84aca392511f1b29a1e54d0171e97d9304af2"} Nov 27 10:40:26 crc kubenswrapper[4806]: I1127 10:40:26.260802 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"794a9a51-b1a0-49c9-8308-d71995e9e23a","Type":"ContainerStarted","Data":"f752a3c5dc361d0c443c9583f70f59fe8ac3cdb8cfeb831ca60ee16d2aa89e3e"} Nov 27 10:40:26 crc kubenswrapper[4806]: I1127 10:40:26.269244 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zhv9q" event={"ID":"9b15a49b-4c6b-4574-a229-25ea19ea2c0f","Type":"ContainerStarted","Data":"f5683037cf95e466d7e550063cda0093b402b439d511b553b012f97b5f32bc23"} Nov 27 10:40:26 crc kubenswrapper[4806]: I1127 10:40:26.269287 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zhv9q" event={"ID":"9b15a49b-4c6b-4574-a229-25ea19ea2c0f","Type":"ContainerStarted","Data":"632d3143bc257d91c8eda74a56404b15576b09b6fdb10e33e1be6dee117544a5"} Nov 27 10:40:26 crc kubenswrapper[4806]: I1127 10:40:26.428545 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-zhv9q" podStartSLOduration=3.428527633 podStartE2EDuration="3.428527633s" podCreationTimestamp="2025-11-27 10:40:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:40:26.340750102 +0000 UTC m=+1130.927340886" watchObservedRunningTime="2025-11-27 10:40:26.428527633 +0000 UTC m=+1131.015118387" Nov 27 10:40:26 crc kubenswrapper[4806]: I1127 10:40:26.455819 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6j7m6"] Nov 27 10:40:27 crc kubenswrapper[4806]: I1127 10:40:27.285276 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6j7m6" event={"ID":"2b0f7898-ee70-46b9-a405-54419ec98b47","Type":"ContainerStarted","Data":"aa917caab719a500e5bc3e742e1ae89c28086ef0cddd38339c19116db015a166"} Nov 27 10:40:27 crc kubenswrapper[4806]: I1127 10:40:27.285778 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6j7m6" event={"ID":"2b0f7898-ee70-46b9-a405-54419ec98b47","Type":"ContainerStarted","Data":"59069674d0cf75651962c69927f2376cf587f853ec32c5eab382df665093a8fb"} Nov 27 10:40:27 crc kubenswrapper[4806]: I1127 10:40:27.294593 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" event={"ID":"d5f6cb41-571f-44d9-b269-33d80225af04","Type":"ContainerStarted","Data":"6c26baad2629ed7504be46a5ea57bf24f86a5a9c89141fd7e88188558240eea9"} Nov 27 10:40:27 crc kubenswrapper[4806]: I1127 10:40:27.307723 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-6j7m6" podStartSLOduration=2.307706654 podStartE2EDuration="2.307706654s" podCreationTimestamp="2025-11-27 10:40:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:40:27.303570552 +0000 UTC m=+1131.890161316" watchObservedRunningTime="2025-11-27 10:40:27.307706654 +0000 UTC m=+1131.894297408" Nov 27 10:40:27 crc kubenswrapper[4806]: I1127 10:40:27.333436 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" podStartSLOduration=3.333415072 podStartE2EDuration="3.333415072s" podCreationTimestamp="2025-11-27 10:40:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:40:27.332935178 +0000 UTC m=+1131.919525952" watchObservedRunningTime="2025-11-27 10:40:27.333415072 +0000 UTC m=+1131.920005836" Nov 27 10:40:28 crc kubenswrapper[4806]: I1127 10:40:28.303080 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:28 crc kubenswrapper[4806]: I1127 10:40:28.512384 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:40:28 crc kubenswrapper[4806]: I1127 10:40:28.531659 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 10:40:30 crc kubenswrapper[4806]: I1127 10:40:30.324563 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2147d816-df60-4343-b25d-901376f1f90f","Type":"ContainerStarted","Data":"8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259"} Nov 27 10:40:30 crc kubenswrapper[4806]: I1127 10:40:30.325549 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2147d816-df60-4343-b25d-901376f1f90f","Type":"ContainerStarted","Data":"f21151b7f346ae2d59590294b1cf3c00b4b379fba406589a60446295c89da865"} Nov 27 10:40:30 crc kubenswrapper[4806]: I1127 10:40:30.325741 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2147d816-df60-4343-b25d-901376f1f90f" containerName="nova-metadata-log" containerID="cri-o://f21151b7f346ae2d59590294b1cf3c00b4b379fba406589a60446295c89da865" gracePeriod=30 Nov 27 10:40:30 crc kubenswrapper[4806]: I1127 10:40:30.325765 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2147d816-df60-4343-b25d-901376f1f90f" containerName="nova-metadata-metadata" containerID="cri-o://8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259" gracePeriod=30 Nov 27 10:40:30 crc kubenswrapper[4806]: I1127 10:40:30.329243 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"794a9a51-b1a0-49c9-8308-d71995e9e23a","Type":"ContainerStarted","Data":"e4e16d74d4eeb1be7714939610a882a402bb7354f011a3cd781fbde78945e693"} Nov 27 10:40:30 crc kubenswrapper[4806]: I1127 10:40:30.329330 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="794a9a51-b1a0-49c9-8308-d71995e9e23a" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e4e16d74d4eeb1be7714939610a882a402bb7354f011a3cd781fbde78945e693" gracePeriod=30 Nov 27 10:40:30 crc kubenswrapper[4806]: I1127 10:40:30.332148 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd","Type":"ContainerStarted","Data":"41613e7e08db95c3ad203ec4bd6f0267f7b901beb9ee4097616eced0aeb7eb2b"} Nov 27 10:40:30 crc kubenswrapper[4806]: I1127 10:40:30.332184 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd","Type":"ContainerStarted","Data":"90e74b818389332af16a1dd514b5c5ae360f310a0e22cc1fe6196237762f3606"} Nov 27 10:40:30 crc kubenswrapper[4806]: I1127 10:40:30.337105 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1363ac1b-7673-4792-82d0-cbccdbd881cb","Type":"ContainerStarted","Data":"89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448"} Nov 27 10:40:30 crc kubenswrapper[4806]: I1127 10:40:30.390773 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.35029354 podStartE2EDuration="6.39073999s" podCreationTimestamp="2025-11-27 10:40:24 +0000 UTC" firstStartedPulling="2025-11-27 10:40:25.475057686 +0000 UTC m=+1130.061648450" lastFinishedPulling="2025-11-27 10:40:29.515504126 +0000 UTC m=+1134.102094900" observedRunningTime="2025-11-27 10:40:30.387505082 +0000 UTC m=+1134.974095856" watchObservedRunningTime="2025-11-27 10:40:30.39073999 +0000 UTC m=+1134.977330754" Nov 27 10:40:30 crc kubenswrapper[4806]: I1127 10:40:30.422289 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.632589407 podStartE2EDuration="6.422260004s" podCreationTimestamp="2025-11-27 10:40:24 +0000 UTC" firstStartedPulling="2025-11-27 10:40:25.722477337 +0000 UTC m=+1130.309068101" lastFinishedPulling="2025-11-27 10:40:29.512147934 +0000 UTC m=+1134.098738698" observedRunningTime="2025-11-27 10:40:30.358222698 +0000 UTC m=+1134.944813462" watchObservedRunningTime="2025-11-27 10:40:30.422260004 +0000 UTC m=+1135.008850768" Nov 27 10:40:30 crc kubenswrapper[4806]: I1127 10:40:30.466277 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.993021793 podStartE2EDuration="6.466223247s" podCreationTimestamp="2025-11-27 10:40:24 +0000 UTC" firstStartedPulling="2025-11-27 10:40:25.026600535 +0000 UTC m=+1129.613191289" lastFinishedPulling="2025-11-27 10:40:29.499801979 +0000 UTC m=+1134.086392743" observedRunningTime="2025-11-27 10:40:30.413645101 +0000 UTC m=+1135.000235875" watchObservedRunningTime="2025-11-27 10:40:30.466223247 +0000 UTC m=+1135.052814011" Nov 27 10:40:30 crc kubenswrapper[4806]: I1127 10:40:30.480527 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.507454774 podStartE2EDuration="6.480498234s" podCreationTimestamp="2025-11-27 10:40:24 +0000 UTC" firstStartedPulling="2025-11-27 10:40:25.541781226 +0000 UTC m=+1130.128371990" lastFinishedPulling="2025-11-27 10:40:29.514824686 +0000 UTC m=+1134.101415450" observedRunningTime="2025-11-27 10:40:30.432635326 +0000 UTC m=+1135.019226090" watchObservedRunningTime="2025-11-27 10:40:30.480498234 +0000 UTC m=+1135.067088998" Nov 27 10:40:31 crc kubenswrapper[4806]: I1127 10:40:31.358819 4806 generic.go:334] "Generic (PLEG): container finished" podID="2147d816-df60-4343-b25d-901376f1f90f" containerID="f21151b7f346ae2d59590294b1cf3c00b4b379fba406589a60446295c89da865" exitCode=143 Nov 27 10:40:31 crc kubenswrapper[4806]: I1127 10:40:31.359394 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2147d816-df60-4343-b25d-901376f1f90f","Type":"ContainerDied","Data":"f21151b7f346ae2d59590294b1cf3c00b4b379fba406589a60446295c89da865"} Nov 27 10:40:34 crc kubenswrapper[4806]: I1127 10:40:34.387901 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 10:40:34 crc kubenswrapper[4806]: I1127 10:40:34.388599 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 10:40:34 crc kubenswrapper[4806]: I1127 10:40:34.713142 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:40:34 crc kubenswrapper[4806]: I1127 10:40:34.755917 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 27 10:40:34 crc kubenswrapper[4806]: I1127 10:40:34.755993 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 27 10:40:34 crc kubenswrapper[4806]: I1127 10:40:34.806471 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 27 10:40:34 crc kubenswrapper[4806]: I1127 10:40:34.887091 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 10:40:34 crc kubenswrapper[4806]: I1127 10:40:34.887177 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 10:40:34 crc kubenswrapper[4806]: I1127 10:40:34.928537 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.003979 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-48ht6"] Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.004208 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58db5546cc-48ht6" podUID="4f7a5d92-530d-4675-9a1b-6cf277064169" containerName="dnsmasq-dns" containerID="cri-o://b9516b0292adbf0cd5fdac0d4a4170075d42dd3fd6674be640f04191a1621962" gracePeriod=10 Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.435835 4806 generic.go:334] "Generic (PLEG): container finished" podID="9b15a49b-4c6b-4574-a229-25ea19ea2c0f" containerID="f5683037cf95e466d7e550063cda0093b402b439d511b553b012f97b5f32bc23" exitCode=0 Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.436075 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zhv9q" event={"ID":"9b15a49b-4c6b-4574-a229-25ea19ea2c0f","Type":"ContainerDied","Data":"f5683037cf95e466d7e550063cda0093b402b439d511b553b012f97b5f32bc23"} Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.445838 4806 generic.go:334] "Generic (PLEG): container finished" podID="4f7a5d92-530d-4675-9a1b-6cf277064169" containerID="b9516b0292adbf0cd5fdac0d4a4170075d42dd3fd6674be640f04191a1621962" exitCode=0 Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.445932 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-48ht6" event={"ID":"4f7a5d92-530d-4675-9a1b-6cf277064169","Type":"ContainerDied","Data":"b9516b0292adbf0cd5fdac0d4a4170075d42dd3fd6674be640f04191a1621962"} Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.450024 4806 generic.go:334] "Generic (PLEG): container finished" podID="2b0f7898-ee70-46b9-a405-54419ec98b47" containerID="aa917caab719a500e5bc3e742e1ae89c28086ef0cddd38339c19116db015a166" exitCode=0 Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.451097 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6j7m6" event={"ID":"2b0f7898-ee70-46b9-a405-54419ec98b47","Type":"ContainerDied","Data":"aa917caab719a500e5bc3e742e1ae89c28086ef0cddd38339c19116db015a166"} Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.481424 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.165:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.481667 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.165:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.521945 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.647802 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.704665 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-config\") pod \"4f7a5d92-530d-4675-9a1b-6cf277064169\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.704772 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-ovsdbserver-nb\") pod \"4f7a5d92-530d-4675-9a1b-6cf277064169\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.704943 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgccm\" (UniqueName: \"kubernetes.io/projected/4f7a5d92-530d-4675-9a1b-6cf277064169-kube-api-access-lgccm\") pod \"4f7a5d92-530d-4675-9a1b-6cf277064169\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.705006 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-ovsdbserver-sb\") pod \"4f7a5d92-530d-4675-9a1b-6cf277064169\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.705031 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-dns-svc\") pod \"4f7a5d92-530d-4675-9a1b-6cf277064169\" (UID: \"4f7a5d92-530d-4675-9a1b-6cf277064169\") " Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.731900 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f7a5d92-530d-4675-9a1b-6cf277064169-kube-api-access-lgccm" (OuterVolumeSpecName: "kube-api-access-lgccm") pod "4f7a5d92-530d-4675-9a1b-6cf277064169" (UID: "4f7a5d92-530d-4675-9a1b-6cf277064169"). InnerVolumeSpecName "kube-api-access-lgccm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.787199 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4f7a5d92-530d-4675-9a1b-6cf277064169" (UID: "4f7a5d92-530d-4675-9a1b-6cf277064169"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.813563 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgccm\" (UniqueName: \"kubernetes.io/projected/4f7a5d92-530d-4675-9a1b-6cf277064169-kube-api-access-lgccm\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.813603 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.822751 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4f7a5d92-530d-4675-9a1b-6cf277064169" (UID: "4f7a5d92-530d-4675-9a1b-6cf277064169"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.830809 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4f7a5d92-530d-4675-9a1b-6cf277064169" (UID: "4f7a5d92-530d-4675-9a1b-6cf277064169"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.846708 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-config" (OuterVolumeSpecName: "config") pod "4f7a5d92-530d-4675-9a1b-6cf277064169" (UID: "4f7a5d92-530d-4675-9a1b-6cf277064169"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.915347 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.915384 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:35 crc kubenswrapper[4806]: I1127 10:40:35.915396 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f7a5d92-530d-4675-9a1b-6cf277064169-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:36 crc kubenswrapper[4806]: I1127 10:40:36.461663 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-48ht6" event={"ID":"4f7a5d92-530d-4675-9a1b-6cf277064169","Type":"ContainerDied","Data":"4ef388f41813a863e716071dae4548acf05379b26d07a2fa2583fcbac1c7cf00"} Nov 27 10:40:36 crc kubenswrapper[4806]: I1127 10:40:36.461752 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-48ht6" Nov 27 10:40:36 crc kubenswrapper[4806]: I1127 10:40:36.461805 4806 scope.go:117] "RemoveContainer" containerID="b9516b0292adbf0cd5fdac0d4a4170075d42dd3fd6674be640f04191a1621962" Nov 27 10:40:36 crc kubenswrapper[4806]: I1127 10:40:36.506392 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-48ht6"] Nov 27 10:40:36 crc kubenswrapper[4806]: I1127 10:40:36.515464 4806 scope.go:117] "RemoveContainer" containerID="41e57d921ac6d6ea0b57795152d1e3c5dcc72a7ad4044801b861fc943187a730" Nov 27 10:40:36 crc kubenswrapper[4806]: I1127 10:40:36.521471 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-48ht6"] Nov 27 10:40:36 crc kubenswrapper[4806]: I1127 10:40:36.965210 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.034592 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.056122 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-config-data\") pod \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.056186 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-scripts\") pod \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.056217 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-config-data\") pod \"2b0f7898-ee70-46b9-a405-54419ec98b47\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.056277 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-scripts\") pod \"2b0f7898-ee70-46b9-a405-54419ec98b47\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.056383 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66xcd\" (UniqueName: \"kubernetes.io/projected/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-kube-api-access-66xcd\") pod \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.056438 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rfld\" (UniqueName: \"kubernetes.io/projected/2b0f7898-ee70-46b9-a405-54419ec98b47-kube-api-access-9rfld\") pod \"2b0f7898-ee70-46b9-a405-54419ec98b47\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.056524 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-combined-ca-bundle\") pod \"2b0f7898-ee70-46b9-a405-54419ec98b47\" (UID: \"2b0f7898-ee70-46b9-a405-54419ec98b47\") " Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.056588 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-combined-ca-bundle\") pod \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\" (UID: \"9b15a49b-4c6b-4574-a229-25ea19ea2c0f\") " Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.067545 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-kube-api-access-66xcd" (OuterVolumeSpecName: "kube-api-access-66xcd") pod "9b15a49b-4c6b-4574-a229-25ea19ea2c0f" (UID: "9b15a49b-4c6b-4574-a229-25ea19ea2c0f"). InnerVolumeSpecName "kube-api-access-66xcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.069340 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-scripts" (OuterVolumeSpecName: "scripts") pod "2b0f7898-ee70-46b9-a405-54419ec98b47" (UID: "2b0f7898-ee70-46b9-a405-54419ec98b47"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.074672 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-scripts" (OuterVolumeSpecName: "scripts") pod "9b15a49b-4c6b-4574-a229-25ea19ea2c0f" (UID: "9b15a49b-4c6b-4574-a229-25ea19ea2c0f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.082931 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b0f7898-ee70-46b9-a405-54419ec98b47-kube-api-access-9rfld" (OuterVolumeSpecName: "kube-api-access-9rfld") pod "2b0f7898-ee70-46b9-a405-54419ec98b47" (UID: "2b0f7898-ee70-46b9-a405-54419ec98b47"). InnerVolumeSpecName "kube-api-access-9rfld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.088504 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b0f7898-ee70-46b9-a405-54419ec98b47" (UID: "2b0f7898-ee70-46b9-a405-54419ec98b47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.094008 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b15a49b-4c6b-4574-a229-25ea19ea2c0f" (UID: "9b15a49b-4c6b-4574-a229-25ea19ea2c0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.099755 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-config-data" (OuterVolumeSpecName: "config-data") pod "2b0f7898-ee70-46b9-a405-54419ec98b47" (UID: "2b0f7898-ee70-46b9-a405-54419ec98b47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.101149 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-config-data" (OuterVolumeSpecName: "config-data") pod "9b15a49b-4c6b-4574-a229-25ea19ea2c0f" (UID: "9b15a49b-4c6b-4574-a229-25ea19ea2c0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.158740 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66xcd\" (UniqueName: \"kubernetes.io/projected/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-kube-api-access-66xcd\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.158819 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rfld\" (UniqueName: \"kubernetes.io/projected/2b0f7898-ee70-46b9-a405-54419ec98b47-kube-api-access-9rfld\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.158833 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.158842 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.158852 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.158863 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b15a49b-4c6b-4574-a229-25ea19ea2c0f-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.158900 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.158911 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b0f7898-ee70-46b9-a405-54419ec98b47-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.479467 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zhv9q" event={"ID":"9b15a49b-4c6b-4574-a229-25ea19ea2c0f","Type":"ContainerDied","Data":"632d3143bc257d91c8eda74a56404b15576b09b6fdb10e33e1be6dee117544a5"} Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.479529 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="632d3143bc257d91c8eda74a56404b15576b09b6fdb10e33e1be6dee117544a5" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.479610 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zhv9q" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.505140 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6j7m6" event={"ID":"2b0f7898-ee70-46b9-a405-54419ec98b47","Type":"ContainerDied","Data":"59069674d0cf75651962c69927f2376cf587f853ec32c5eab382df665093a8fb"} Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.505247 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59069674d0cf75651962c69927f2376cf587f853ec32c5eab382df665093a8fb" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.505257 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6j7m6" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.592434 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 27 10:40:37 crc kubenswrapper[4806]: E1127 10:40:37.592951 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b15a49b-4c6b-4574-a229-25ea19ea2c0f" containerName="nova-manage" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.592974 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b15a49b-4c6b-4574-a229-25ea19ea2c0f" containerName="nova-manage" Nov 27 10:40:37 crc kubenswrapper[4806]: E1127 10:40:37.592989 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f7a5d92-530d-4675-9a1b-6cf277064169" containerName="dnsmasq-dns" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.592996 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f7a5d92-530d-4675-9a1b-6cf277064169" containerName="dnsmasq-dns" Nov 27 10:40:37 crc kubenswrapper[4806]: E1127 10:40:37.593007 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f7a5d92-530d-4675-9a1b-6cf277064169" containerName="init" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.593014 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f7a5d92-530d-4675-9a1b-6cf277064169" containerName="init" Nov 27 10:40:37 crc kubenswrapper[4806]: E1127 10:40:37.593034 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b0f7898-ee70-46b9-a405-54419ec98b47" containerName="nova-cell1-conductor-db-sync" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.593042 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b0f7898-ee70-46b9-a405-54419ec98b47" containerName="nova-cell1-conductor-db-sync" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.593268 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b15a49b-4c6b-4574-a229-25ea19ea2c0f" containerName="nova-manage" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.593289 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b0f7898-ee70-46b9-a405-54419ec98b47" containerName="nova-cell1-conductor-db-sync" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.593295 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f7a5d92-530d-4675-9a1b-6cf277064169" containerName="dnsmasq-dns" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.595722 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.601191 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.669389 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d93051aa-7a15-40df-92cb-7e9cef6103a8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d93051aa-7a15-40df-92cb-7e9cef6103a8\") " pod="openstack/nova-cell1-conductor-0" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.669594 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d93051aa-7a15-40df-92cb-7e9cef6103a8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d93051aa-7a15-40df-92cb-7e9cef6103a8\") " pod="openstack/nova-cell1-conductor-0" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.669639 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vntj\" (UniqueName: \"kubernetes.io/projected/d93051aa-7a15-40df-92cb-7e9cef6103a8-kube-api-access-4vntj\") pod \"nova-cell1-conductor-0\" (UID: \"d93051aa-7a15-40df-92cb-7e9cef6103a8\") " pod="openstack/nova-cell1-conductor-0" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.673610 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.738123 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.738495 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" containerName="nova-api-log" containerID="cri-o://90e74b818389332af16a1dd514b5c5ae360f310a0e22cc1fe6196237762f3606" gracePeriod=30 Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.739304 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" containerName="nova-api-api" containerID="cri-o://41613e7e08db95c3ad203ec4bd6f0267f7b901beb9ee4097616eced0aeb7eb2b" gracePeriod=30 Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.774397 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d93051aa-7a15-40df-92cb-7e9cef6103a8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d93051aa-7a15-40df-92cb-7e9cef6103a8\") " pod="openstack/nova-cell1-conductor-0" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.774645 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d93051aa-7a15-40df-92cb-7e9cef6103a8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d93051aa-7a15-40df-92cb-7e9cef6103a8\") " pod="openstack/nova-cell1-conductor-0" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.775664 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vntj\" (UniqueName: \"kubernetes.io/projected/d93051aa-7a15-40df-92cb-7e9cef6103a8-kube-api-access-4vntj\") pod \"nova-cell1-conductor-0\" (UID: \"d93051aa-7a15-40df-92cb-7e9cef6103a8\") " pod="openstack/nova-cell1-conductor-0" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.777792 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.778052 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1363ac1b-7673-4792-82d0-cbccdbd881cb" containerName="nova-scheduler-scheduler" containerID="cri-o://89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448" gracePeriod=30 Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.799461 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d93051aa-7a15-40df-92cb-7e9cef6103a8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d93051aa-7a15-40df-92cb-7e9cef6103a8\") " pod="openstack/nova-cell1-conductor-0" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.809867 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vntj\" (UniqueName: \"kubernetes.io/projected/d93051aa-7a15-40df-92cb-7e9cef6103a8-kube-api-access-4vntj\") pod \"nova-cell1-conductor-0\" (UID: \"d93051aa-7a15-40df-92cb-7e9cef6103a8\") " pod="openstack/nova-cell1-conductor-0" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.819385 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d93051aa-7a15-40df-92cb-7e9cef6103a8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d93051aa-7a15-40df-92cb-7e9cef6103a8\") " pod="openstack/nova-cell1-conductor-0" Nov 27 10:40:37 crc kubenswrapper[4806]: I1127 10:40:37.938018 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 27 10:40:38 crc kubenswrapper[4806]: I1127 10:40:38.140567 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f7a5d92-530d-4675-9a1b-6cf277064169" path="/var/lib/kubelet/pods/4f7a5d92-530d-4675-9a1b-6cf277064169/volumes" Nov 27 10:40:38 crc kubenswrapper[4806]: I1127 10:40:38.318515 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 27 10:40:38 crc kubenswrapper[4806]: I1127 10:40:38.494633 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 27 10:40:38 crc kubenswrapper[4806]: I1127 10:40:38.516526 4806 generic.go:334] "Generic (PLEG): container finished" podID="3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" containerID="90e74b818389332af16a1dd514b5c5ae360f310a0e22cc1fe6196237762f3606" exitCode=143 Nov 27 10:40:38 crc kubenswrapper[4806]: I1127 10:40:38.516811 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd","Type":"ContainerDied","Data":"90e74b818389332af16a1dd514b5c5ae360f310a0e22cc1fe6196237762f3606"} Nov 27 10:40:39 crc kubenswrapper[4806]: I1127 10:40:39.529411 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d93051aa-7a15-40df-92cb-7e9cef6103a8","Type":"ContainerStarted","Data":"ee03c78c21422cb63dca612acf7c45a4c1782a22253caf315cf6a806ab9320a9"} Nov 27 10:40:39 crc kubenswrapper[4806]: I1127 10:40:39.529698 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d93051aa-7a15-40df-92cb-7e9cef6103a8","Type":"ContainerStarted","Data":"57208c476eaaad0903ace032708b3fb70d3ba19187de606d04329a7cb4dc6157"} Nov 27 10:40:39 crc kubenswrapper[4806]: I1127 10:40:39.530146 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 27 10:40:39 crc kubenswrapper[4806]: I1127 10:40:39.556410 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.556389883 podStartE2EDuration="2.556389883s" podCreationTimestamp="2025-11-27 10:40:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:40:39.549578258 +0000 UTC m=+1144.136169012" watchObservedRunningTime="2025-11-27 10:40:39.556389883 +0000 UTC m=+1144.142980657" Nov 27 10:40:39 crc kubenswrapper[4806]: E1127 10:40:39.758376 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 27 10:40:39 crc kubenswrapper[4806]: E1127 10:40:39.760531 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 27 10:40:39 crc kubenswrapper[4806]: E1127 10:40:39.762448 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 27 10:40:39 crc kubenswrapper[4806]: E1127 10:40:39.762486 4806 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1363ac1b-7673-4792-82d0-cbccdbd881cb" containerName="nova-scheduler-scheduler" Nov 27 10:40:41 crc kubenswrapper[4806]: I1127 10:40:41.518352 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 10:40:41 crc kubenswrapper[4806]: I1127 10:40:41.520714 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="beb345b2-15fd-4db1-8ea0-c862b7820fc0" containerName="kube-state-metrics" containerID="cri-o://1356dc8aaca88cf9d159fd1ff7cd38621ef0d9cca9817734724a264b1475f593" gracePeriod=30 Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.106620 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.282272 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kh2rj\" (UniqueName: \"kubernetes.io/projected/beb345b2-15fd-4db1-8ea0-c862b7820fc0-kube-api-access-kh2rj\") pod \"beb345b2-15fd-4db1-8ea0-c862b7820fc0\" (UID: \"beb345b2-15fd-4db1-8ea0-c862b7820fc0\") " Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.322122 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beb345b2-15fd-4db1-8ea0-c862b7820fc0-kube-api-access-kh2rj" (OuterVolumeSpecName: "kube-api-access-kh2rj") pod "beb345b2-15fd-4db1-8ea0-c862b7820fc0" (UID: "beb345b2-15fd-4db1-8ea0-c862b7820fc0"). InnerVolumeSpecName "kube-api-access-kh2rj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.385453 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kh2rj\" (UniqueName: \"kubernetes.io/projected/beb345b2-15fd-4db1-8ea0-c862b7820fc0-kube-api-access-kh2rj\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.462119 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.577802 4806 generic.go:334] "Generic (PLEG): container finished" podID="3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" containerID="41613e7e08db95c3ad203ec4bd6f0267f7b901beb9ee4097616eced0aeb7eb2b" exitCode=0 Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.579117 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd","Type":"ContainerDied","Data":"41613e7e08db95c3ad203ec4bd6f0267f7b901beb9ee4097616eced0aeb7eb2b"} Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.581169 4806 generic.go:334] "Generic (PLEG): container finished" podID="beb345b2-15fd-4db1-8ea0-c862b7820fc0" containerID="1356dc8aaca88cf9d159fd1ff7cd38621ef0d9cca9817734724a264b1475f593" exitCode=2 Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.581332 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"beb345b2-15fd-4db1-8ea0-c862b7820fc0","Type":"ContainerDied","Data":"1356dc8aaca88cf9d159fd1ff7cd38621ef0d9cca9817734724a264b1475f593"} Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.581427 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"beb345b2-15fd-4db1-8ea0-c862b7820fc0","Type":"ContainerDied","Data":"be3f12a2906eee11239ad63bf0c5aa8de6cb1041aa34c4da893b37649eadd8aa"} Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.581518 4806 scope.go:117] "RemoveContainer" containerID="1356dc8aaca88cf9d159fd1ff7cd38621ef0d9cca9817734724a264b1475f593" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.581799 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.589274 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1363ac1b-7673-4792-82d0-cbccdbd881cb-config-data\") pod \"1363ac1b-7673-4792-82d0-cbccdbd881cb\" (UID: \"1363ac1b-7673-4792-82d0-cbccdbd881cb\") " Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.589495 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1363ac1b-7673-4792-82d0-cbccdbd881cb-combined-ca-bundle\") pod \"1363ac1b-7673-4792-82d0-cbccdbd881cb\" (UID: \"1363ac1b-7673-4792-82d0-cbccdbd881cb\") " Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.589580 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wxrm\" (UniqueName: \"kubernetes.io/projected/1363ac1b-7673-4792-82d0-cbccdbd881cb-kube-api-access-5wxrm\") pod \"1363ac1b-7673-4792-82d0-cbccdbd881cb\" (UID: \"1363ac1b-7673-4792-82d0-cbccdbd881cb\") " Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.594300 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1363ac1b-7673-4792-82d0-cbccdbd881cb-kube-api-access-5wxrm" (OuterVolumeSpecName: "kube-api-access-5wxrm") pod "1363ac1b-7673-4792-82d0-cbccdbd881cb" (UID: "1363ac1b-7673-4792-82d0-cbccdbd881cb"). InnerVolumeSpecName "kube-api-access-5wxrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.598112 4806 generic.go:334] "Generic (PLEG): container finished" podID="1363ac1b-7673-4792-82d0-cbccdbd881cb" containerID="89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448" exitCode=0 Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.598178 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1363ac1b-7673-4792-82d0-cbccdbd881cb","Type":"ContainerDied","Data":"89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448"} Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.599122 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1363ac1b-7673-4792-82d0-cbccdbd881cb","Type":"ContainerDied","Data":"4cd538a9ca86ada617893f51760d29e64903bf77826030ffe83f31452f8fb114"} Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.599285 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.632046 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1363ac1b-7673-4792-82d0-cbccdbd881cb-config-data" (OuterVolumeSpecName: "config-data") pod "1363ac1b-7673-4792-82d0-cbccdbd881cb" (UID: "1363ac1b-7673-4792-82d0-cbccdbd881cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.639545 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.647781 4806 scope.go:117] "RemoveContainer" containerID="1356dc8aaca88cf9d159fd1ff7cd38621ef0d9cca9817734724a264b1475f593" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.650823 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1363ac1b-7673-4792-82d0-cbccdbd881cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1363ac1b-7673-4792-82d0-cbccdbd881cb" (UID: "1363ac1b-7673-4792-82d0-cbccdbd881cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:42 crc kubenswrapper[4806]: E1127 10:40:42.650832 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1356dc8aaca88cf9d159fd1ff7cd38621ef0d9cca9817734724a264b1475f593\": container with ID starting with 1356dc8aaca88cf9d159fd1ff7cd38621ef0d9cca9817734724a264b1475f593 not found: ID does not exist" containerID="1356dc8aaca88cf9d159fd1ff7cd38621ef0d9cca9817734724a264b1475f593" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.650913 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1356dc8aaca88cf9d159fd1ff7cd38621ef0d9cca9817734724a264b1475f593"} err="failed to get container status \"1356dc8aaca88cf9d159fd1ff7cd38621ef0d9cca9817734724a264b1475f593\": rpc error: code = NotFound desc = could not find container \"1356dc8aaca88cf9d159fd1ff7cd38621ef0d9cca9817734724a264b1475f593\": container with ID starting with 1356dc8aaca88cf9d159fd1ff7cd38621ef0d9cca9817734724a264b1475f593 not found: ID does not exist" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.650966 4806 scope.go:117] "RemoveContainer" containerID="89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.680370 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.694459 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1363ac1b-7673-4792-82d0-cbccdbd881cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.694500 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wxrm\" (UniqueName: \"kubernetes.io/projected/1363ac1b-7673-4792-82d0-cbccdbd881cb-kube-api-access-5wxrm\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.694515 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1363ac1b-7673-4792-82d0-cbccdbd881cb-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.696513 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 10:40:42 crc kubenswrapper[4806]: E1127 10:40:42.697103 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1363ac1b-7673-4792-82d0-cbccdbd881cb" containerName="nova-scheduler-scheduler" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.697131 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="1363ac1b-7673-4792-82d0-cbccdbd881cb" containerName="nova-scheduler-scheduler" Nov 27 10:40:42 crc kubenswrapper[4806]: E1127 10:40:42.697182 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb345b2-15fd-4db1-8ea0-c862b7820fc0" containerName="kube-state-metrics" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.697193 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb345b2-15fd-4db1-8ea0-c862b7820fc0" containerName="kube-state-metrics" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.697442 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="1363ac1b-7673-4792-82d0-cbccdbd881cb" containerName="nova-scheduler-scheduler" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.697490 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="beb345b2-15fd-4db1-8ea0-c862b7820fc0" containerName="kube-state-metrics" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.698415 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.702890 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.703630 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.741297 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.770061 4806 scope.go:117] "RemoveContainer" containerID="89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448" Nov 27 10:40:42 crc kubenswrapper[4806]: E1127 10:40:42.772582 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448\": container with ID starting with 89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448 not found: ID does not exist" containerID="89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.772654 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448"} err="failed to get container status \"89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448\": rpc error: code = NotFound desc = could not find container \"89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448\": container with ID starting with 89409436bbdbd0dba97f4566f97a364e810706fe50599df8dd028c94fd52d448 not found: ID does not exist" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.800054 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053fa4b1-ca2c-4e00-b0ad-1d076e853dec-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"053fa4b1-ca2c-4e00-b0ad-1d076e853dec\") " pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.805335 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/053fa4b1-ca2c-4e00-b0ad-1d076e853dec-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"053fa4b1-ca2c-4e00-b0ad-1d076e853dec\") " pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.805463 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/053fa4b1-ca2c-4e00-b0ad-1d076e853dec-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"053fa4b1-ca2c-4e00-b0ad-1d076e853dec\") " pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.805603 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlrqr\" (UniqueName: \"kubernetes.io/projected/053fa4b1-ca2c-4e00-b0ad-1d076e853dec-kube-api-access-dlrqr\") pod \"kube-state-metrics-0\" (UID: \"053fa4b1-ca2c-4e00-b0ad-1d076e853dec\") " pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.907615 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/053fa4b1-ca2c-4e00-b0ad-1d076e853dec-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"053fa4b1-ca2c-4e00-b0ad-1d076e853dec\") " pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.907673 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/053fa4b1-ca2c-4e00-b0ad-1d076e853dec-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"053fa4b1-ca2c-4e00-b0ad-1d076e853dec\") " pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.907731 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlrqr\" (UniqueName: \"kubernetes.io/projected/053fa4b1-ca2c-4e00-b0ad-1d076e853dec-kube-api-access-dlrqr\") pod \"kube-state-metrics-0\" (UID: \"053fa4b1-ca2c-4e00-b0ad-1d076e853dec\") " pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.907801 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053fa4b1-ca2c-4e00-b0ad-1d076e853dec-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"053fa4b1-ca2c-4e00-b0ad-1d076e853dec\") " pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.915395 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053fa4b1-ca2c-4e00-b0ad-1d076e853dec-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"053fa4b1-ca2c-4e00-b0ad-1d076e853dec\") " pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.920948 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/053fa4b1-ca2c-4e00-b0ad-1d076e853dec-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"053fa4b1-ca2c-4e00-b0ad-1d076e853dec\") " pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.922332 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/053fa4b1-ca2c-4e00-b0ad-1d076e853dec-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"053fa4b1-ca2c-4e00-b0ad-1d076e853dec\") " pod="openstack/kube-state-metrics-0" Nov 27 10:40:42 crc kubenswrapper[4806]: I1127 10:40:42.936018 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlrqr\" (UniqueName: \"kubernetes.io/projected/053fa4b1-ca2c-4e00-b0ad-1d076e853dec-kube-api-access-dlrqr\") pod \"kube-state-metrics-0\" (UID: \"053fa4b1-ca2c-4e00-b0ad-1d076e853dec\") " pod="openstack/kube-state-metrics-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.016702 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.032567 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.040435 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.067434 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:40:43 crc kubenswrapper[4806]: E1127 10:40:43.068148 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" containerName="nova-api-api" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.068253 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" containerName="nova-api-api" Nov 27 10:40:43 crc kubenswrapper[4806]: E1127 10:40:43.068331 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" containerName="nova-api-log" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.068381 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" containerName="nova-api-log" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.068671 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" containerName="nova-api-api" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.068778 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" containerName="nova-api-log" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.069466 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.072174 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.083246 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.088039 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.230929 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-logs\") pod \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.231095 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nnv2\" (UniqueName: \"kubernetes.io/projected/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-kube-api-access-9nnv2\") pod \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.231157 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-config-data\") pod \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.231266 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-combined-ca-bundle\") pod \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\" (UID: \"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd\") " Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.231676 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dec9511-dd0f-4f87-922a-4959333ec7fc-config-data\") pod \"nova-scheduler-0\" (UID: \"8dec9511-dd0f-4f87-922a-4959333ec7fc\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.231697 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dec9511-dd0f-4f87-922a-4959333ec7fc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8dec9511-dd0f-4f87-922a-4959333ec7fc\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.231797 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9bvm\" (UniqueName: \"kubernetes.io/projected/8dec9511-dd0f-4f87-922a-4959333ec7fc-kube-api-access-z9bvm\") pod \"nova-scheduler-0\" (UID: \"8dec9511-dd0f-4f87-922a-4959333ec7fc\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.234965 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-logs" (OuterVolumeSpecName: "logs") pod "3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" (UID: "3ccbc1ed-f210-4db1-a8a4-aa022a9913dd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.257623 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-kube-api-access-9nnv2" (OuterVolumeSpecName: "kube-api-access-9nnv2") pod "3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" (UID: "3ccbc1ed-f210-4db1-a8a4-aa022a9913dd"). InnerVolumeSpecName "kube-api-access-9nnv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.287059 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" (UID: "3ccbc1ed-f210-4db1-a8a4-aa022a9913dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.294518 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-config-data" (OuterVolumeSpecName: "config-data") pod "3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" (UID: "3ccbc1ed-f210-4db1-a8a4-aa022a9913dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.333884 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dec9511-dd0f-4f87-922a-4959333ec7fc-config-data\") pod \"nova-scheduler-0\" (UID: \"8dec9511-dd0f-4f87-922a-4959333ec7fc\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.333972 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dec9511-dd0f-4f87-922a-4959333ec7fc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8dec9511-dd0f-4f87-922a-4959333ec7fc\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.334058 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9bvm\" (UniqueName: \"kubernetes.io/projected/8dec9511-dd0f-4f87-922a-4959333ec7fc-kube-api-access-z9bvm\") pod \"nova-scheduler-0\" (UID: \"8dec9511-dd0f-4f87-922a-4959333ec7fc\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.334492 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.335986 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-logs\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.336076 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nnv2\" (UniqueName: \"kubernetes.io/projected/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-kube-api-access-9nnv2\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.336098 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.338711 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dec9511-dd0f-4f87-922a-4959333ec7fc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8dec9511-dd0f-4f87-922a-4959333ec7fc\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.346178 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dec9511-dd0f-4f87-922a-4959333ec7fc-config-data\") pod \"nova-scheduler-0\" (UID: \"8dec9511-dd0f-4f87-922a-4959333ec7fc\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.354892 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9bvm\" (UniqueName: \"kubernetes.io/projected/8dec9511-dd0f-4f87-922a-4959333ec7fc-kube-api-access-z9bvm\") pod \"nova-scheduler-0\" (UID: \"8dec9511-dd0f-4f87-922a-4959333ec7fc\") " pod="openstack/nova-scheduler-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.399922 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.616447 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.616451 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3ccbc1ed-f210-4db1-a8a4-aa022a9913dd","Type":"ContainerDied","Data":"9b398ac377e4aee8cf9478a41e03a8484f9f026c3d14977902ce08a681f32f3e"} Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.617210 4806 scope.go:117] "RemoveContainer" containerID="41613e7e08db95c3ad203ec4bd6f0267f7b901beb9ee4097616eced0aeb7eb2b" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.641595 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.643296 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="ceilometer-central-agent" containerID="cri-o://0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164" gracePeriod=30 Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.643523 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="proxy-httpd" containerID="cri-o://d3f52d60ec4e75a0974a6d083f8492ad4a50d8e1228ff9e15a20355a9c5bf5ef" gracePeriod=30 Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.643654 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="sg-core" containerID="cri-o://b5f0194541e5bb8ceccb08213ca8d9a563c56563f016c629db20ff57047f9dda" gracePeriod=30 Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.644733 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="ceilometer-notification-agent" containerID="cri-o://e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97" gracePeriod=30 Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.675424 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.766475 4806 scope.go:117] "RemoveContainer" containerID="90e74b818389332af16a1dd514b5c5ae360f310a0e22cc1fe6196237762f3606" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.811353 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.827568 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.843488 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.845095 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.853647 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.856005 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.954047 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54c143ac-14ac-45b1-b5a4-0b945167827f-config-data\") pod \"nova-api-0\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " pod="openstack/nova-api-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.954152 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54c143ac-14ac-45b1-b5a4-0b945167827f-logs\") pod \"nova-api-0\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " pod="openstack/nova-api-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.954338 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jml7x\" (UniqueName: \"kubernetes.io/projected/54c143ac-14ac-45b1-b5a4-0b945167827f-kube-api-access-jml7x\") pod \"nova-api-0\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " pod="openstack/nova-api-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.954459 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c143ac-14ac-45b1-b5a4-0b945167827f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " pod="openstack/nova-api-0" Nov 27 10:40:43 crc kubenswrapper[4806]: I1127 10:40:43.994158 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.057815 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54c143ac-14ac-45b1-b5a4-0b945167827f-config-data\") pod \"nova-api-0\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " pod="openstack/nova-api-0" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.058586 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54c143ac-14ac-45b1-b5a4-0b945167827f-logs\") pod \"nova-api-0\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " pod="openstack/nova-api-0" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.059005 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54c143ac-14ac-45b1-b5a4-0b945167827f-logs\") pod \"nova-api-0\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " pod="openstack/nova-api-0" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.059414 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jml7x\" (UniqueName: \"kubernetes.io/projected/54c143ac-14ac-45b1-b5a4-0b945167827f-kube-api-access-jml7x\") pod \"nova-api-0\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " pod="openstack/nova-api-0" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.059789 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c143ac-14ac-45b1-b5a4-0b945167827f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " pod="openstack/nova-api-0" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.062312 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54c143ac-14ac-45b1-b5a4-0b945167827f-config-data\") pod \"nova-api-0\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " pod="openstack/nova-api-0" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.065067 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c143ac-14ac-45b1-b5a4-0b945167827f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " pod="openstack/nova-api-0" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.078836 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jml7x\" (UniqueName: \"kubernetes.io/projected/54c143ac-14ac-45b1-b5a4-0b945167827f-kube-api-access-jml7x\") pod \"nova-api-0\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " pod="openstack/nova-api-0" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.139247 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1363ac1b-7673-4792-82d0-cbccdbd881cb" path="/var/lib/kubelet/pods/1363ac1b-7673-4792-82d0-cbccdbd881cb/volumes" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.140273 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ccbc1ed-f210-4db1-a8a4-aa022a9913dd" path="/var/lib/kubelet/pods/3ccbc1ed-f210-4db1-a8a4-aa022a9913dd/volumes" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.143328 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beb345b2-15fd-4db1-8ea0-c862b7820fc0" path="/var/lib/kubelet/pods/beb345b2-15fd-4db1-8ea0-c862b7820fc0/volumes" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.222442 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.652717 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"053fa4b1-ca2c-4e00-b0ad-1d076e853dec","Type":"ContainerStarted","Data":"fcbb2756548f6e69a7eaac374b92b756cbdbc56a497c699b8908ca034febdeb0"} Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.653644 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"053fa4b1-ca2c-4e00-b0ad-1d076e853dec","Type":"ContainerStarted","Data":"71210773332e3d12cadc38c591c65f31eb1d4538d353a26259dd39789dd1b49a"} Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.653732 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.657768 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8dec9511-dd0f-4f87-922a-4959333ec7fc","Type":"ContainerStarted","Data":"2777010d2ceb747b0fccc30eda6c208196fdba9745e3774019ab2abed1cbb058"} Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.657799 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8dec9511-dd0f-4f87-922a-4959333ec7fc","Type":"ContainerStarted","Data":"26eb377c96c711ab0a449442efa0101f9158aa199ffc366b5a5825ad17e243c0"} Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.663876 4806 generic.go:334] "Generic (PLEG): container finished" podID="3c820267-b03d-4918-b7fd-aa30d9205034" containerID="d3f52d60ec4e75a0974a6d083f8492ad4a50d8e1228ff9e15a20355a9c5bf5ef" exitCode=0 Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.663908 4806 generic.go:334] "Generic (PLEG): container finished" podID="3c820267-b03d-4918-b7fd-aa30d9205034" containerID="b5f0194541e5bb8ceccb08213ca8d9a563c56563f016c629db20ff57047f9dda" exitCode=2 Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.663918 4806 generic.go:334] "Generic (PLEG): container finished" podID="3c820267-b03d-4918-b7fd-aa30d9205034" containerID="0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164" exitCode=0 Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.663955 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c820267-b03d-4918-b7fd-aa30d9205034","Type":"ContainerDied","Data":"d3f52d60ec4e75a0974a6d083f8492ad4a50d8e1228ff9e15a20355a9c5bf5ef"} Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.663977 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c820267-b03d-4918-b7fd-aa30d9205034","Type":"ContainerDied","Data":"b5f0194541e5bb8ceccb08213ca8d9a563c56563f016c629db20ff57047f9dda"} Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.663987 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c820267-b03d-4918-b7fd-aa30d9205034","Type":"ContainerDied","Data":"0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164"} Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.680017 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.232544022 podStartE2EDuration="2.679995055s" podCreationTimestamp="2025-11-27 10:40:42 +0000 UTC" firstStartedPulling="2025-11-27 10:40:43.766321378 +0000 UTC m=+1148.352912142" lastFinishedPulling="2025-11-27 10:40:44.213772411 +0000 UTC m=+1148.800363175" observedRunningTime="2025-11-27 10:40:44.67467336 +0000 UTC m=+1149.261264124" watchObservedRunningTime="2025-11-27 10:40:44.679995055 +0000 UTC m=+1149.266585819" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.705384 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.705356173 podStartE2EDuration="1.705356173s" podCreationTimestamp="2025-11-27 10:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:40:44.693256014 +0000 UTC m=+1149.279846778" watchObservedRunningTime="2025-11-27 10:40:44.705356173 +0000 UTC m=+1149.291946937" Nov 27 10:40:44 crc kubenswrapper[4806]: I1127 10:40:44.752658 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:40:45 crc kubenswrapper[4806]: I1127 10:40:45.682997 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"54c143ac-14ac-45b1-b5a4-0b945167827f","Type":"ContainerStarted","Data":"dba5b7ad175a67c8def72c72279199594a907a8e3a2971e5561d17483fd5b35a"} Nov 27 10:40:45 crc kubenswrapper[4806]: I1127 10:40:45.683464 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"54c143ac-14ac-45b1-b5a4-0b945167827f","Type":"ContainerStarted","Data":"29c90a7a98e9c0091ec3359bc9822b0da9b54c07f5f7d2a2832d6fba9fb36904"} Nov 27 10:40:45 crc kubenswrapper[4806]: I1127 10:40:45.683479 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"54c143ac-14ac-45b1-b5a4-0b945167827f","Type":"ContainerStarted","Data":"4ea8ffedf962b26e83609a7c716e5a4b36137243ae5473b6b9a4a9895d802c08"} Nov 27 10:40:45 crc kubenswrapper[4806]: I1127 10:40:45.715695 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.715652269 podStartE2EDuration="2.715652269s" podCreationTimestamp="2025-11-27 10:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:40:45.706139371 +0000 UTC m=+1150.292730135" watchObservedRunningTime="2025-11-27 10:40:45.715652269 +0000 UTC m=+1150.302243033" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.397896 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.540626 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c820267-b03d-4918-b7fd-aa30d9205034-log-httpd\") pod \"3c820267-b03d-4918-b7fd-aa30d9205034\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.540827 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-sg-core-conf-yaml\") pod \"3c820267-b03d-4918-b7fd-aa30d9205034\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.540938 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c820267-b03d-4918-b7fd-aa30d9205034-run-httpd\") pod \"3c820267-b03d-4918-b7fd-aa30d9205034\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.541379 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c820267-b03d-4918-b7fd-aa30d9205034-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3c820267-b03d-4918-b7fd-aa30d9205034" (UID: "3c820267-b03d-4918-b7fd-aa30d9205034"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.541439 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c820267-b03d-4918-b7fd-aa30d9205034-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3c820267-b03d-4918-b7fd-aa30d9205034" (UID: "3c820267-b03d-4918-b7fd-aa30d9205034"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.541062 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvsv4\" (UniqueName: \"kubernetes.io/projected/3c820267-b03d-4918-b7fd-aa30d9205034-kube-api-access-xvsv4\") pod \"3c820267-b03d-4918-b7fd-aa30d9205034\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.542224 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-config-data\") pod \"3c820267-b03d-4918-b7fd-aa30d9205034\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.542297 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-scripts\") pod \"3c820267-b03d-4918-b7fd-aa30d9205034\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.542419 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-combined-ca-bundle\") pod \"3c820267-b03d-4918-b7fd-aa30d9205034\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.549424 4806 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c820267-b03d-4918-b7fd-aa30d9205034-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.549466 4806 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c820267-b03d-4918-b7fd-aa30d9205034-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.549753 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c820267-b03d-4918-b7fd-aa30d9205034-kube-api-access-xvsv4" (OuterVolumeSpecName: "kube-api-access-xvsv4") pod "3c820267-b03d-4918-b7fd-aa30d9205034" (UID: "3c820267-b03d-4918-b7fd-aa30d9205034"). InnerVolumeSpecName "kube-api-access-xvsv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.563346 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-scripts" (OuterVolumeSpecName: "scripts") pod "3c820267-b03d-4918-b7fd-aa30d9205034" (UID: "3c820267-b03d-4918-b7fd-aa30d9205034"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.593350 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3c820267-b03d-4918-b7fd-aa30d9205034" (UID: "3c820267-b03d-4918-b7fd-aa30d9205034"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.651647 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.651700 4806 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.651717 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvsv4\" (UniqueName: \"kubernetes.io/projected/3c820267-b03d-4918-b7fd-aa30d9205034-kube-api-access-xvsv4\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:47 crc kubenswrapper[4806]: E1127 10:40:47.679552 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-combined-ca-bundle podName:3c820267-b03d-4918-b7fd-aa30d9205034 nodeName:}" failed. No retries permitted until 2025-11-27 10:40:48.179505575 +0000 UTC m=+1152.766096339 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-combined-ca-bundle") pod "3c820267-b03d-4918-b7fd-aa30d9205034" (UID: "3c820267-b03d-4918-b7fd-aa30d9205034") : error deleting /var/lib/kubelet/pods/3c820267-b03d-4918-b7fd-aa30d9205034/volume-subpaths: remove /var/lib/kubelet/pods/3c820267-b03d-4918-b7fd-aa30d9205034/volume-subpaths: no such file or directory Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.682318 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-config-data" (OuterVolumeSpecName: "config-data") pod "3c820267-b03d-4918-b7fd-aa30d9205034" (UID: "3c820267-b03d-4918-b7fd-aa30d9205034"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.711161 4806 generic.go:334] "Generic (PLEG): container finished" podID="3c820267-b03d-4918-b7fd-aa30d9205034" containerID="e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97" exitCode=0 Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.711225 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c820267-b03d-4918-b7fd-aa30d9205034","Type":"ContainerDied","Data":"e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97"} Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.711312 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c820267-b03d-4918-b7fd-aa30d9205034","Type":"ContainerDied","Data":"b019ca168fcb15082699af8505faab7b40002d6538a63c94c573df8a23a9ebc7"} Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.711333 4806 scope.go:117] "RemoveContainer" containerID="d3f52d60ec4e75a0974a6d083f8492ad4a50d8e1228ff9e15a20355a9c5bf5ef" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.711692 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.733066 4806 scope.go:117] "RemoveContainer" containerID="b5f0194541e5bb8ceccb08213ca8d9a563c56563f016c629db20ff57047f9dda" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.754098 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.754568 4806 scope.go:117] "RemoveContainer" containerID="e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.776043 4806 scope.go:117] "RemoveContainer" containerID="0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.798191 4806 scope.go:117] "RemoveContainer" containerID="d3f52d60ec4e75a0974a6d083f8492ad4a50d8e1228ff9e15a20355a9c5bf5ef" Nov 27 10:40:47 crc kubenswrapper[4806]: E1127 10:40:47.798756 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3f52d60ec4e75a0974a6d083f8492ad4a50d8e1228ff9e15a20355a9c5bf5ef\": container with ID starting with d3f52d60ec4e75a0974a6d083f8492ad4a50d8e1228ff9e15a20355a9c5bf5ef not found: ID does not exist" containerID="d3f52d60ec4e75a0974a6d083f8492ad4a50d8e1228ff9e15a20355a9c5bf5ef" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.798833 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3f52d60ec4e75a0974a6d083f8492ad4a50d8e1228ff9e15a20355a9c5bf5ef"} err="failed to get container status \"d3f52d60ec4e75a0974a6d083f8492ad4a50d8e1228ff9e15a20355a9c5bf5ef\": rpc error: code = NotFound desc = could not find container \"d3f52d60ec4e75a0974a6d083f8492ad4a50d8e1228ff9e15a20355a9c5bf5ef\": container with ID starting with d3f52d60ec4e75a0974a6d083f8492ad4a50d8e1228ff9e15a20355a9c5bf5ef not found: ID does not exist" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.798879 4806 scope.go:117] "RemoveContainer" containerID="b5f0194541e5bb8ceccb08213ca8d9a563c56563f016c629db20ff57047f9dda" Nov 27 10:40:47 crc kubenswrapper[4806]: E1127 10:40:47.799283 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5f0194541e5bb8ceccb08213ca8d9a563c56563f016c629db20ff57047f9dda\": container with ID starting with b5f0194541e5bb8ceccb08213ca8d9a563c56563f016c629db20ff57047f9dda not found: ID does not exist" containerID="b5f0194541e5bb8ceccb08213ca8d9a563c56563f016c629db20ff57047f9dda" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.799330 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5f0194541e5bb8ceccb08213ca8d9a563c56563f016c629db20ff57047f9dda"} err="failed to get container status \"b5f0194541e5bb8ceccb08213ca8d9a563c56563f016c629db20ff57047f9dda\": rpc error: code = NotFound desc = could not find container \"b5f0194541e5bb8ceccb08213ca8d9a563c56563f016c629db20ff57047f9dda\": container with ID starting with b5f0194541e5bb8ceccb08213ca8d9a563c56563f016c629db20ff57047f9dda not found: ID does not exist" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.799362 4806 scope.go:117] "RemoveContainer" containerID="e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97" Nov 27 10:40:47 crc kubenswrapper[4806]: E1127 10:40:47.799690 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97\": container with ID starting with e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97 not found: ID does not exist" containerID="e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.799721 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97"} err="failed to get container status \"e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97\": rpc error: code = NotFound desc = could not find container \"e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97\": container with ID starting with e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97 not found: ID does not exist" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.799741 4806 scope.go:117] "RemoveContainer" containerID="0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164" Nov 27 10:40:47 crc kubenswrapper[4806]: E1127 10:40:47.800344 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164\": container with ID starting with 0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164 not found: ID does not exist" containerID="0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.800419 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164"} err="failed to get container status \"0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164\": rpc error: code = NotFound desc = could not find container \"0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164\": container with ID starting with 0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164 not found: ID does not exist" Nov 27 10:40:47 crc kubenswrapper[4806]: I1127 10:40:47.987717 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.263522 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-combined-ca-bundle\") pod \"3c820267-b03d-4918-b7fd-aa30d9205034\" (UID: \"3c820267-b03d-4918-b7fd-aa30d9205034\") " Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.276980 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c820267-b03d-4918-b7fd-aa30d9205034" (UID: "3c820267-b03d-4918-b7fd-aa30d9205034"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.358385 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.367374 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c820267-b03d-4918-b7fd-aa30d9205034-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.382079 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.395277 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:40:48 crc kubenswrapper[4806]: E1127 10:40:48.395903 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="ceilometer-central-agent" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.395972 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="ceilometer-central-agent" Nov 27 10:40:48 crc kubenswrapper[4806]: E1127 10:40:48.396062 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="ceilometer-notification-agent" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.396113 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="ceilometer-notification-agent" Nov 27 10:40:48 crc kubenswrapper[4806]: E1127 10:40:48.396183 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="sg-core" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.396268 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="sg-core" Nov 27 10:40:48 crc kubenswrapper[4806]: E1127 10:40:48.396426 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="proxy-httpd" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.396496 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="proxy-httpd" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.396723 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="proxy-httpd" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.396806 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="sg-core" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.396859 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="ceilometer-notification-agent" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.396914 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" containerName="ceilometer-central-agent" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.398883 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.402310 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.405559 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.405748 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.405842 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.408603 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.572916 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-config-data\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.573363 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.573493 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.573683 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgrdt\" (UniqueName: \"kubernetes.io/projected/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-kube-api-access-xgrdt\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.573786 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.573898 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-run-httpd\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.574068 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-log-httpd\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.574176 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-scripts\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.676672 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-run-httpd\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.676761 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-log-httpd\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.676788 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-scripts\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.676826 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-config-data\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.676858 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.676895 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.676992 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgrdt\" (UniqueName: \"kubernetes.io/projected/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-kube-api-access-xgrdt\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.677025 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.677269 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-run-httpd\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.678135 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-log-httpd\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.683293 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.683877 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.685763 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.692455 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-config-data\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.696469 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-scripts\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.704662 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgrdt\" (UniqueName: \"kubernetes.io/projected/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-kube-api-access-xgrdt\") pod \"ceilometer-0\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " pod="openstack/ceilometer-0" Nov 27 10:40:48 crc kubenswrapper[4806]: I1127 10:40:48.755828 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:40:49 crc kubenswrapper[4806]: I1127 10:40:49.456936 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:40:49 crc kubenswrapper[4806]: I1127 10:40:49.735797 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94c3d68b-963f-4ba6-b32f-dbbdcaefd377","Type":"ContainerStarted","Data":"735c0d4098393563b1f8f5ca747cf7329c22dfa4382e6609b8f1a3679ee2be48"} Nov 27 10:40:50 crc kubenswrapper[4806]: I1127 10:40:50.128062 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c820267-b03d-4918-b7fd-aa30d9205034" path="/var/lib/kubelet/pods/3c820267-b03d-4918-b7fd-aa30d9205034/volumes" Nov 27 10:40:50 crc kubenswrapper[4806]: I1127 10:40:50.746376 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94c3d68b-963f-4ba6-b32f-dbbdcaefd377","Type":"ContainerStarted","Data":"0c495272c04e41dccc4c83ccc21e0d26e6d3c18550936024a641225651d6ce0c"} Nov 27 10:40:51 crc kubenswrapper[4806]: I1127 10:40:51.758098 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94c3d68b-963f-4ba6-b32f-dbbdcaefd377","Type":"ContainerStarted","Data":"1511d764c4cd9652da460a249acec5c3439d3acbb23725f75f84b121beee3b3b"} Nov 27 10:40:51 crc kubenswrapper[4806]: I1127 10:40:51.759078 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94c3d68b-963f-4ba6-b32f-dbbdcaefd377","Type":"ContainerStarted","Data":"0a07048e6b536ac07e6408d46557d83e52d17aae162d8589c4c18fbd247f6b34"} Nov 27 10:40:53 crc kubenswrapper[4806]: I1127 10:40:53.102223 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 27 10:40:53 crc kubenswrapper[4806]: I1127 10:40:53.400677 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 27 10:40:53 crc kubenswrapper[4806]: I1127 10:40:53.435249 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 27 10:40:53 crc kubenswrapper[4806]: I1127 10:40:53.778972 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94c3d68b-963f-4ba6-b32f-dbbdcaefd377","Type":"ContainerStarted","Data":"5b571fd767f00ba6ae6b79935a45e658bc009d075a4c472d94a4df45b3648d7a"} Nov 27 10:40:53 crc kubenswrapper[4806]: I1127 10:40:53.822310 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.851102054 podStartE2EDuration="5.822290554s" podCreationTimestamp="2025-11-27 10:40:48 +0000 UTC" firstStartedPulling="2025-11-27 10:40:49.481593394 +0000 UTC m=+1154.068184158" lastFinishedPulling="2025-11-27 10:40:53.452781894 +0000 UTC m=+1158.039372658" observedRunningTime="2025-11-27 10:40:53.809761705 +0000 UTC m=+1158.396352469" watchObservedRunningTime="2025-11-27 10:40:53.822290554 +0000 UTC m=+1158.408881318" Nov 27 10:40:53 crc kubenswrapper[4806]: I1127 10:40:53.823522 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 27 10:40:54 crc kubenswrapper[4806]: I1127 10:40:54.223085 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 10:40:54 crc kubenswrapper[4806]: I1127 10:40:54.223483 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 10:40:54 crc kubenswrapper[4806]: I1127 10:40:54.788480 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 10:40:55 crc kubenswrapper[4806]: I1127 10:40:55.305479 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="54c143ac-14ac-45b1-b5a4-0b945167827f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.174:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 10:40:55 crc kubenswrapper[4806]: I1127 10:40:55.305522 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="54c143ac-14ac-45b1-b5a4-0b945167827f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.174:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 27 10:40:59 crc kubenswrapper[4806]: E1127 10:40:59.587778 4806 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/d5db4914c9ee10f67453915b42be9d64cd5d9a43eda097c9c4766faafd3ac325/diff" to get inode usage: stat /var/lib/containers/storage/overlay/d5db4914c9ee10f67453915b42be9d64cd5d9a43eda097c9c4766faafd3ac325/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_kube-state-metrics-0_beb345b2-15fd-4db1-8ea0-c862b7820fc0/kube-state-metrics/0.log" to get inode usage: stat /var/log/pods/openstack_kube-state-metrics-0_beb345b2-15fd-4db1-8ea0-c862b7820fc0/kube-state-metrics/0.log: no such file or directory Nov 27 10:41:00 crc kubenswrapper[4806]: E1127 10:41:00.389059 4806 manager.go:1116] Failed to create existing container: /kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ccbc1ed_f210_4db1_a8a4_aa022a9913dd.slice/crio-9b398ac377e4aee8cf9478a41e03a8484f9f026c3d14977902ce08a681f32f3e: Error finding container 9b398ac377e4aee8cf9478a41e03a8484f9f026c3d14977902ce08a681f32f3e: Status 404 returned error can't find the container with id 9b398ac377e4aee8cf9478a41e03a8484f9f026c3d14977902ce08a681f32f3e Nov 27 10:41:00 crc kubenswrapper[4806]: E1127 10:41:00.737033 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod794a9a51_b1a0_49c9_8308_d71995e9e23a.slice/crio-e4e16d74d4eeb1be7714939610a882a402bb7354f011a3cd781fbde78945e693.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c820267_b03d_4918_b7fd_aa30d9205034.slice/crio-e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2147d816_df60_4343_b25d_901376f1f90f.slice/crio-8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c820267_b03d_4918_b7fd_aa30d9205034.slice/crio-b019ca168fcb15082699af8505faab7b40002d6538a63c94c573df8a23a9ebc7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c820267_b03d_4918_b7fd_aa30d9205034.slice/crio-conmon-0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c820267_b03d_4918_b7fd_aa30d9205034.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c820267_b03d_4918_b7fd_aa30d9205034.slice/crio-0ace0ae4015e0e048fab2b2575bf4f15d683a4d04e4c50e1cfa00a159ad22164.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod794a9a51_b1a0_49c9_8308_d71995e9e23a.slice/crio-conmon-e4e16d74d4eeb1be7714939610a882a402bb7354f011a3cd781fbde78945e693.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c820267_b03d_4918_b7fd_aa30d9205034.slice/crio-conmon-e8fa6ecb1c46147212a5838cf080b5bbeb401b46d0ce82530be75547704a6e97.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2147d816_df60_4343_b25d_901376f1f90f.slice/crio-conmon-8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259.scope\": RecentStats: unable to find data in memory cache]" Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.853826 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.903299 4806 generic.go:334] "Generic (PLEG): container finished" podID="2147d816-df60-4343-b25d-901376f1f90f" containerID="8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259" exitCode=137 Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.903633 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2147d816-df60-4343-b25d-901376f1f90f","Type":"ContainerDied","Data":"8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259"} Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.903661 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2147d816-df60-4343-b25d-901376f1f90f","Type":"ContainerDied","Data":"59ceed370c930105930798f821e84aca392511f1b29a1e54d0171e97d9304af2"} Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.903679 4806 scope.go:117] "RemoveContainer" containerID="8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259" Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.903789 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.920152 4806 generic.go:334] "Generic (PLEG): container finished" podID="794a9a51-b1a0-49c9-8308-d71995e9e23a" containerID="e4e16d74d4eeb1be7714939610a882a402bb7354f011a3cd781fbde78945e693" exitCode=137 Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.920203 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"794a9a51-b1a0-49c9-8308-d71995e9e23a","Type":"ContainerDied","Data":"e4e16d74d4eeb1be7714939610a882a402bb7354f011a3cd781fbde78945e693"} Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.920248 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"794a9a51-b1a0-49c9-8308-d71995e9e23a","Type":"ContainerDied","Data":"f752a3c5dc361d0c443c9583f70f59fe8ac3cdb8cfeb831ca60ee16d2aa89e3e"} Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.920261 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f752a3c5dc361d0c443c9583f70f59fe8ac3cdb8cfeb831ca60ee16d2aa89e3e" Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.949867 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2147d816-df60-4343-b25d-901376f1f90f-logs\") pod \"2147d816-df60-4343-b25d-901376f1f90f\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.949985 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2147d816-df60-4343-b25d-901376f1f90f-config-data\") pod \"2147d816-df60-4343-b25d-901376f1f90f\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.950020 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h7sn\" (UniqueName: \"kubernetes.io/projected/2147d816-df60-4343-b25d-901376f1f90f-kube-api-access-5h7sn\") pod \"2147d816-df60-4343-b25d-901376f1f90f\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.950086 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2147d816-df60-4343-b25d-901376f1f90f-combined-ca-bundle\") pod \"2147d816-df60-4343-b25d-901376f1f90f\" (UID: \"2147d816-df60-4343-b25d-901376f1f90f\") " Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.950306 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2147d816-df60-4343-b25d-901376f1f90f-logs" (OuterVolumeSpecName: "logs") pod "2147d816-df60-4343-b25d-901376f1f90f" (UID: "2147d816-df60-4343-b25d-901376f1f90f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.950759 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2147d816-df60-4343-b25d-901376f1f90f-logs\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.957451 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2147d816-df60-4343-b25d-901376f1f90f-kube-api-access-5h7sn" (OuterVolumeSpecName: "kube-api-access-5h7sn") pod "2147d816-df60-4343-b25d-901376f1f90f" (UID: "2147d816-df60-4343-b25d-901376f1f90f"). InnerVolumeSpecName "kube-api-access-5h7sn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.961740 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.986629 4806 scope.go:117] "RemoveContainer" containerID="f21151b7f346ae2d59590294b1cf3c00b4b379fba406589a60446295c89da865" Nov 27 10:41:00 crc kubenswrapper[4806]: I1127 10:41:00.993822 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2147d816-df60-4343-b25d-901376f1f90f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2147d816-df60-4343-b25d-901376f1f90f" (UID: "2147d816-df60-4343-b25d-901376f1f90f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.005934 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2147d816-df60-4343-b25d-901376f1f90f-config-data" (OuterVolumeSpecName: "config-data") pod "2147d816-df60-4343-b25d-901376f1f90f" (UID: "2147d816-df60-4343-b25d-901376f1f90f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.018181 4806 scope.go:117] "RemoveContainer" containerID="8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259" Nov 27 10:41:01 crc kubenswrapper[4806]: E1127 10:41:01.018675 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259\": container with ID starting with 8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259 not found: ID does not exist" containerID="8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.018723 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259"} err="failed to get container status \"8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259\": rpc error: code = NotFound desc = could not find container \"8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259\": container with ID starting with 8b08c766d0c872037678bd4149a0e412cc0c4fa59d00a18f11b0ca6941be0259 not found: ID does not exist" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.018754 4806 scope.go:117] "RemoveContainer" containerID="f21151b7f346ae2d59590294b1cf3c00b4b379fba406589a60446295c89da865" Nov 27 10:41:01 crc kubenswrapper[4806]: E1127 10:41:01.018961 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f21151b7f346ae2d59590294b1cf3c00b4b379fba406589a60446295c89da865\": container with ID starting with f21151b7f346ae2d59590294b1cf3c00b4b379fba406589a60446295c89da865 not found: ID does not exist" containerID="f21151b7f346ae2d59590294b1cf3c00b4b379fba406589a60446295c89da865" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.018987 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21151b7f346ae2d59590294b1cf3c00b4b379fba406589a60446295c89da865"} err="failed to get container status \"f21151b7f346ae2d59590294b1cf3c00b4b379fba406589a60446295c89da865\": rpc error: code = NotFound desc = could not find container \"f21151b7f346ae2d59590294b1cf3c00b4b379fba406589a60446295c89da865\": container with ID starting with f21151b7f346ae2d59590294b1cf3c00b4b379fba406589a60446295c89da865 not found: ID does not exist" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.052503 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794a9a51-b1a0-49c9-8308-d71995e9e23a-combined-ca-bundle\") pod \"794a9a51-b1a0-49c9-8308-d71995e9e23a\" (UID: \"794a9a51-b1a0-49c9-8308-d71995e9e23a\") " Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.052666 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5xjq\" (UniqueName: \"kubernetes.io/projected/794a9a51-b1a0-49c9-8308-d71995e9e23a-kube-api-access-q5xjq\") pod \"794a9a51-b1a0-49c9-8308-d71995e9e23a\" (UID: \"794a9a51-b1a0-49c9-8308-d71995e9e23a\") " Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.053164 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794a9a51-b1a0-49c9-8308-d71995e9e23a-config-data\") pod \"794a9a51-b1a0-49c9-8308-d71995e9e23a\" (UID: \"794a9a51-b1a0-49c9-8308-d71995e9e23a\") " Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.053714 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2147d816-df60-4343-b25d-901376f1f90f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.053735 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h7sn\" (UniqueName: \"kubernetes.io/projected/2147d816-df60-4343-b25d-901376f1f90f-kube-api-access-5h7sn\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.053752 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2147d816-df60-4343-b25d-901376f1f90f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.056934 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/794a9a51-b1a0-49c9-8308-d71995e9e23a-kube-api-access-q5xjq" (OuterVolumeSpecName: "kube-api-access-q5xjq") pod "794a9a51-b1a0-49c9-8308-d71995e9e23a" (UID: "794a9a51-b1a0-49c9-8308-d71995e9e23a"). InnerVolumeSpecName "kube-api-access-q5xjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.077045 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/794a9a51-b1a0-49c9-8308-d71995e9e23a-config-data" (OuterVolumeSpecName: "config-data") pod "794a9a51-b1a0-49c9-8308-d71995e9e23a" (UID: "794a9a51-b1a0-49c9-8308-d71995e9e23a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.082899 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/794a9a51-b1a0-49c9-8308-d71995e9e23a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "794a9a51-b1a0-49c9-8308-d71995e9e23a" (UID: "794a9a51-b1a0-49c9-8308-d71995e9e23a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.155511 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5xjq\" (UniqueName: \"kubernetes.io/projected/794a9a51-b1a0-49c9-8308-d71995e9e23a-kube-api-access-q5xjq\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.155553 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794a9a51-b1a0-49c9-8308-d71995e9e23a-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.155565 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794a9a51-b1a0-49c9-8308-d71995e9e23a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.244294 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.254796 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.279889 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:41:01 crc kubenswrapper[4806]: E1127 10:41:01.280320 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="794a9a51-b1a0-49c9-8308-d71995e9e23a" containerName="nova-cell1-novncproxy-novncproxy" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.280336 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="794a9a51-b1a0-49c9-8308-d71995e9e23a" containerName="nova-cell1-novncproxy-novncproxy" Nov 27 10:41:01 crc kubenswrapper[4806]: E1127 10:41:01.280354 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2147d816-df60-4343-b25d-901376f1f90f" containerName="nova-metadata-metadata" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.280362 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2147d816-df60-4343-b25d-901376f1f90f" containerName="nova-metadata-metadata" Nov 27 10:41:01 crc kubenswrapper[4806]: E1127 10:41:01.280392 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2147d816-df60-4343-b25d-901376f1f90f" containerName="nova-metadata-log" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.280403 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2147d816-df60-4343-b25d-901376f1f90f" containerName="nova-metadata-log" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.280644 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="794a9a51-b1a0-49c9-8308-d71995e9e23a" containerName="nova-cell1-novncproxy-novncproxy" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.280674 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="2147d816-df60-4343-b25d-901376f1f90f" containerName="nova-metadata-metadata" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.280686 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="2147d816-df60-4343-b25d-901376f1f90f" containerName="nova-metadata-log" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.282418 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.301398 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.303958 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.306343 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.360390 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.360913 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.360995 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f339550-0757-4afd-b60b-3eacf7d20b4f-logs\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.361195 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwm2t\" (UniqueName: \"kubernetes.io/projected/9f339550-0757-4afd-b60b-3eacf7d20b4f-kube-api-access-xwm2t\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.361409 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-config-data\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.465201 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.465391 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.465511 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f339550-0757-4afd-b60b-3eacf7d20b4f-logs\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.465595 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwm2t\" (UniqueName: \"kubernetes.io/projected/9f339550-0757-4afd-b60b-3eacf7d20b4f-kube-api-access-xwm2t\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.465701 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-config-data\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.465988 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f339550-0757-4afd-b60b-3eacf7d20b4f-logs\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.469988 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.470070 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-config-data\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.480015 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.488213 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwm2t\" (UniqueName: \"kubernetes.io/projected/9f339550-0757-4afd-b60b-3eacf7d20b4f-kube-api-access-xwm2t\") pod \"nova-metadata-0\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.601640 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 10:41:01 crc kubenswrapper[4806]: I1127 10:41:01.931771 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.001963 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.025568 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.037011 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.038783 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.042536 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.043052 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.044516 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.047661 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.129394 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2147d816-df60-4343-b25d-901376f1f90f" path="/var/lib/kubelet/pods/2147d816-df60-4343-b25d-901376f1f90f/volumes" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.130000 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="794a9a51-b1a0-49c9-8308-d71995e9e23a" path="/var/lib/kubelet/pods/794a9a51-b1a0-49c9-8308-d71995e9e23a/volumes" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.130676 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.181516 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa061b40-cf89-478c-a82e-957785f57c69-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.182608 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa061b40-cf89-478c-a82e-957785f57c69-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.182666 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv725\" (UniqueName: \"kubernetes.io/projected/aa061b40-cf89-478c-a82e-957785f57c69-kube-api-access-bv725\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.182754 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa061b40-cf89-478c-a82e-957785f57c69-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.183127 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa061b40-cf89-478c-a82e-957785f57c69-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.285706 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa061b40-cf89-478c-a82e-957785f57c69-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.286067 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv725\" (UniqueName: \"kubernetes.io/projected/aa061b40-cf89-478c-a82e-957785f57c69-kube-api-access-bv725\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.286179 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa061b40-cf89-478c-a82e-957785f57c69-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.286365 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa061b40-cf89-478c-a82e-957785f57c69-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.286528 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa061b40-cf89-478c-a82e-957785f57c69-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.292774 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa061b40-cf89-478c-a82e-957785f57c69-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.294250 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa061b40-cf89-478c-a82e-957785f57c69-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.298379 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa061b40-cf89-478c-a82e-957785f57c69-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.299410 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa061b40-cf89-478c-a82e-957785f57c69-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.317332 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv725\" (UniqueName: \"kubernetes.io/projected/aa061b40-cf89-478c-a82e-957785f57c69-kube-api-access-bv725\") pod \"nova-cell1-novncproxy-0\" (UID: \"aa061b40-cf89-478c-a82e-957785f57c69\") " pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.379920 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.883724 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 27 10:41:02 crc kubenswrapper[4806]: W1127 10:41:02.889857 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa061b40_cf89_478c_a82e_957785f57c69.slice/crio-9b7a751fd17b3398109e02300d6ee9eec05c80c2c93ecc94f8dc7ab63144e764 WatchSource:0}: Error finding container 9b7a751fd17b3398109e02300d6ee9eec05c80c2c93ecc94f8dc7ab63144e764: Status 404 returned error can't find the container with id 9b7a751fd17b3398109e02300d6ee9eec05c80c2c93ecc94f8dc7ab63144e764 Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.957516 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"aa061b40-cf89-478c-a82e-957785f57c69","Type":"ContainerStarted","Data":"9b7a751fd17b3398109e02300d6ee9eec05c80c2c93ecc94f8dc7ab63144e764"} Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.961919 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f339550-0757-4afd-b60b-3eacf7d20b4f","Type":"ContainerStarted","Data":"1cdc33f7b797971f6e784bcc81bf09e3cffba7ae86c819edc5255c4ce55ae7d7"} Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.962088 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f339550-0757-4afd-b60b-3eacf7d20b4f","Type":"ContainerStarted","Data":"0606ae3db73a2db520cbdf66a9662e26d8384e49ed46a484add7af18dae70a47"} Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.962287 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f339550-0757-4afd-b60b-3eacf7d20b4f","Type":"ContainerStarted","Data":"79f4280f26a1b8f84024f89cc0b89ea49019626eec5cd01efb48d2eedcc21c96"} Nov 27 10:41:02 crc kubenswrapper[4806]: I1127 10:41:02.994134 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.994099775 podStartE2EDuration="1.994099775s" podCreationTimestamp="2025-11-27 10:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:41:02.993725495 +0000 UTC m=+1167.580316259" watchObservedRunningTime="2025-11-27 10:41:02.994099775 +0000 UTC m=+1167.580690539" Nov 27 10:41:03 crc kubenswrapper[4806]: I1127 10:41:03.972343 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"aa061b40-cf89-478c-a82e-957785f57c69","Type":"ContainerStarted","Data":"4d660f09a677cba622fc10055842e5921db49ead1cc0e75d365c7e0091da4246"} Nov 27 10:41:04 crc kubenswrapper[4806]: I1127 10:41:04.011739 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.01170998 podStartE2EDuration="3.01170998s" podCreationTimestamp="2025-11-27 10:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:41:03.995384138 +0000 UTC m=+1168.581974912" watchObservedRunningTime="2025-11-27 10:41:04.01170998 +0000 UTC m=+1168.598300744" Nov 27 10:41:04 crc kubenswrapper[4806]: I1127 10:41:04.228804 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 27 10:41:04 crc kubenswrapper[4806]: I1127 10:41:04.229467 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 27 10:41:04 crc kubenswrapper[4806]: I1127 10:41:04.232915 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 27 10:41:04 crc kubenswrapper[4806]: I1127 10:41:04.238468 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 27 10:41:04 crc kubenswrapper[4806]: I1127 10:41:04.983367 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 27 10:41:04 crc kubenswrapper[4806]: I1127 10:41:04.987957 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.300990 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-dkkxb"] Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.311787 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.321396 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-dkkxb"] Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.368265 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-config\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.368426 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.368487 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.368527 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.368569 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npm4s\" (UniqueName: \"kubernetes.io/projected/356b8268-4acb-4ccd-92c5-b085c9fdd884-kube-api-access-npm4s\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.477018 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.477371 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.477427 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npm4s\" (UniqueName: \"kubernetes.io/projected/356b8268-4acb-4ccd-92c5-b085c9fdd884-kube-api-access-npm4s\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.477452 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-config\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.477540 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.478349 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.478534 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.478960 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-config\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.479145 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.527337 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npm4s\" (UniqueName: \"kubernetes.io/projected/356b8268-4acb-4ccd-92c5-b085c9fdd884-kube-api-access-npm4s\") pod \"dnsmasq-dns-68d4b6d797-dkkxb\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:05 crc kubenswrapper[4806]: I1127 10:41:05.646635 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:06 crc kubenswrapper[4806]: I1127 10:41:06.309989 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-dkkxb"] Nov 27 10:41:06 crc kubenswrapper[4806]: W1127 10:41:06.325618 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod356b8268_4acb_4ccd_92c5_b085c9fdd884.slice/crio-78a6503c474989813ccb7e4362e29231fb130d7c88c2187e25ecbf348c2c6368 WatchSource:0}: Error finding container 78a6503c474989813ccb7e4362e29231fb130d7c88c2187e25ecbf348c2c6368: Status 404 returned error can't find the container with id 78a6503c474989813ccb7e4362e29231fb130d7c88c2187e25ecbf348c2c6368 Nov 27 10:41:06 crc kubenswrapper[4806]: I1127 10:41:06.604049 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 10:41:06 crc kubenswrapper[4806]: I1127 10:41:06.604620 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 10:41:07 crc kubenswrapper[4806]: I1127 10:41:07.014580 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" event={"ID":"356b8268-4acb-4ccd-92c5-b085c9fdd884","Type":"ContainerDied","Data":"4a6dd3d7cdc811fe1d7ab3a7da7da2163bec8f67f20aa03b9775b8e638e66db7"} Nov 27 10:41:07 crc kubenswrapper[4806]: I1127 10:41:07.014492 4806 generic.go:334] "Generic (PLEG): container finished" podID="356b8268-4acb-4ccd-92c5-b085c9fdd884" containerID="4a6dd3d7cdc811fe1d7ab3a7da7da2163bec8f67f20aa03b9775b8e638e66db7" exitCode=0 Nov 27 10:41:07 crc kubenswrapper[4806]: I1127 10:41:07.016192 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" event={"ID":"356b8268-4acb-4ccd-92c5-b085c9fdd884","Type":"ContainerStarted","Data":"78a6503c474989813ccb7e4362e29231fb130d7c88c2187e25ecbf348c2c6368"} Nov 27 10:41:07 crc kubenswrapper[4806]: I1127 10:41:07.380825 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:08 crc kubenswrapper[4806]: I1127 10:41:08.030791 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" event={"ID":"356b8268-4acb-4ccd-92c5-b085c9fdd884","Type":"ContainerStarted","Data":"14fde9f23ade1454eddf0c593d445bad7df3fa72e3f58ec0d0de0fc4368be60c"} Nov 27 10:41:08 crc kubenswrapper[4806]: I1127 10:41:08.030958 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:08 crc kubenswrapper[4806]: I1127 10:41:08.059536 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" podStartSLOduration=3.0594708 podStartE2EDuration="3.0594708s" podCreationTimestamp="2025-11-27 10:41:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:41:08.050583539 +0000 UTC m=+1172.637174313" watchObservedRunningTime="2025-11-27 10:41:08.0594708 +0000 UTC m=+1172.646061694" Nov 27 10:41:08 crc kubenswrapper[4806]: I1127 10:41:08.442655 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:41:08 crc kubenswrapper[4806]: I1127 10:41:08.450138 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="54c143ac-14ac-45b1-b5a4-0b945167827f" containerName="nova-api-log" containerID="cri-o://29c90a7a98e9c0091ec3359bc9822b0da9b54c07f5f7d2a2832d6fba9fb36904" gracePeriod=30 Nov 27 10:41:08 crc kubenswrapper[4806]: I1127 10:41:08.450455 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="54c143ac-14ac-45b1-b5a4-0b945167827f" containerName="nova-api-api" containerID="cri-o://dba5b7ad175a67c8def72c72279199594a907a8e3a2971e5561d17483fd5b35a" gracePeriod=30 Nov 27 10:41:08 crc kubenswrapper[4806]: I1127 10:41:08.915671 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:41:08 crc kubenswrapper[4806]: I1127 10:41:08.916516 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="ceilometer-central-agent" containerID="cri-o://0c495272c04e41dccc4c83ccc21e0d26e6d3c18550936024a641225651d6ce0c" gracePeriod=30 Nov 27 10:41:08 crc kubenswrapper[4806]: I1127 10:41:08.916586 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="sg-core" containerID="cri-o://1511d764c4cd9652da460a249acec5c3439d3acbb23725f75f84b121beee3b3b" gracePeriod=30 Nov 27 10:41:08 crc kubenswrapper[4806]: I1127 10:41:08.916623 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="ceilometer-notification-agent" containerID="cri-o://0a07048e6b536ac07e6408d46557d83e52d17aae162d8589c4c18fbd247f6b34" gracePeriod=30 Nov 27 10:41:08 crc kubenswrapper[4806]: I1127 10:41:08.916632 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="proxy-httpd" containerID="cri-o://5b571fd767f00ba6ae6b79935a45e658bc009d075a4c472d94a4df45b3648d7a" gracePeriod=30 Nov 27 10:41:09 crc kubenswrapper[4806]: I1127 10:41:09.022599 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.175:3000/\": read tcp 10.217.0.2:34326->10.217.0.175:3000: read: connection reset by peer" Nov 27 10:41:09 crc kubenswrapper[4806]: I1127 10:41:09.060456 4806 generic.go:334] "Generic (PLEG): container finished" podID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerID="1511d764c4cd9652da460a249acec5c3439d3acbb23725f75f84b121beee3b3b" exitCode=2 Nov 27 10:41:09 crc kubenswrapper[4806]: I1127 10:41:09.060882 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94c3d68b-963f-4ba6-b32f-dbbdcaefd377","Type":"ContainerDied","Data":"1511d764c4cd9652da460a249acec5c3439d3acbb23725f75f84b121beee3b3b"} Nov 27 10:41:09 crc kubenswrapper[4806]: I1127 10:41:09.064626 4806 generic.go:334] "Generic (PLEG): container finished" podID="54c143ac-14ac-45b1-b5a4-0b945167827f" containerID="29c90a7a98e9c0091ec3359bc9822b0da9b54c07f5f7d2a2832d6fba9fb36904" exitCode=143 Nov 27 10:41:09 crc kubenswrapper[4806]: I1127 10:41:09.065407 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"54c143ac-14ac-45b1-b5a4-0b945167827f","Type":"ContainerDied","Data":"29c90a7a98e9c0091ec3359bc9822b0da9b54c07f5f7d2a2832d6fba9fb36904"} Nov 27 10:41:10 crc kubenswrapper[4806]: I1127 10:41:10.079675 4806 generic.go:334] "Generic (PLEG): container finished" podID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerID="5b571fd767f00ba6ae6b79935a45e658bc009d075a4c472d94a4df45b3648d7a" exitCode=0 Nov 27 10:41:10 crc kubenswrapper[4806]: I1127 10:41:10.080898 4806 generic.go:334] "Generic (PLEG): container finished" podID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerID="0c495272c04e41dccc4c83ccc21e0d26e6d3c18550936024a641225651d6ce0c" exitCode=0 Nov 27 10:41:10 crc kubenswrapper[4806]: I1127 10:41:10.079823 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94c3d68b-963f-4ba6-b32f-dbbdcaefd377","Type":"ContainerDied","Data":"5b571fd767f00ba6ae6b79935a45e658bc009d075a4c472d94a4df45b3648d7a"} Nov 27 10:41:10 crc kubenswrapper[4806]: I1127 10:41:10.081158 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94c3d68b-963f-4ba6-b32f-dbbdcaefd377","Type":"ContainerDied","Data":"0c495272c04e41dccc4c83ccc21e0d26e6d3c18550936024a641225651d6ce0c"} Nov 27 10:41:11 crc kubenswrapper[4806]: I1127 10:41:11.603349 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 27 10:41:11 crc kubenswrapper[4806]: I1127 10:41:11.604309 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.110952 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.112894 4806 generic.go:334] "Generic (PLEG): container finished" podID="54c143ac-14ac-45b1-b5a4-0b945167827f" containerID="dba5b7ad175a67c8def72c72279199594a907a8e3a2971e5561d17483fd5b35a" exitCode=0 Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.112950 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"54c143ac-14ac-45b1-b5a4-0b945167827f","Type":"ContainerDied","Data":"dba5b7ad175a67c8def72c72279199594a907a8e3a2971e5561d17483fd5b35a"} Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.112987 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"54c143ac-14ac-45b1-b5a4-0b945167827f","Type":"ContainerDied","Data":"4ea8ffedf962b26e83609a7c716e5a4b36137243ae5473b6b9a4a9895d802c08"} Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.113005 4806 scope.go:117] "RemoveContainer" containerID="dba5b7ad175a67c8def72c72279199594a907a8e3a2971e5561d17483fd5b35a" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.142936 4806 scope.go:117] "RemoveContainer" containerID="29c90a7a98e9c0091ec3359bc9822b0da9b54c07f5f7d2a2832d6fba9fb36904" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.152391 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54c143ac-14ac-45b1-b5a4-0b945167827f-logs\") pod \"54c143ac-14ac-45b1-b5a4-0b945167827f\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.152625 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c143ac-14ac-45b1-b5a4-0b945167827f-combined-ca-bundle\") pod \"54c143ac-14ac-45b1-b5a4-0b945167827f\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.152979 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jml7x\" (UniqueName: \"kubernetes.io/projected/54c143ac-14ac-45b1-b5a4-0b945167827f-kube-api-access-jml7x\") pod \"54c143ac-14ac-45b1-b5a4-0b945167827f\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.153178 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54c143ac-14ac-45b1-b5a4-0b945167827f-config-data\") pod \"54c143ac-14ac-45b1-b5a4-0b945167827f\" (UID: \"54c143ac-14ac-45b1-b5a4-0b945167827f\") " Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.153386 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54c143ac-14ac-45b1-b5a4-0b945167827f-logs" (OuterVolumeSpecName: "logs") pod "54c143ac-14ac-45b1-b5a4-0b945167827f" (UID: "54c143ac-14ac-45b1-b5a4-0b945167827f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.153914 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54c143ac-14ac-45b1-b5a4-0b945167827f-logs\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.160969 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54c143ac-14ac-45b1-b5a4-0b945167827f-kube-api-access-jml7x" (OuterVolumeSpecName: "kube-api-access-jml7x") pod "54c143ac-14ac-45b1-b5a4-0b945167827f" (UID: "54c143ac-14ac-45b1-b5a4-0b945167827f"). InnerVolumeSpecName "kube-api-access-jml7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.210698 4806 scope.go:117] "RemoveContainer" containerID="dba5b7ad175a67c8def72c72279199594a907a8e3a2971e5561d17483fd5b35a" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.215403 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54c143ac-14ac-45b1-b5a4-0b945167827f-config-data" (OuterVolumeSpecName: "config-data") pod "54c143ac-14ac-45b1-b5a4-0b945167827f" (UID: "54c143ac-14ac-45b1-b5a4-0b945167827f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:12 crc kubenswrapper[4806]: E1127 10:41:12.221577 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dba5b7ad175a67c8def72c72279199594a907a8e3a2971e5561d17483fd5b35a\": container with ID starting with dba5b7ad175a67c8def72c72279199594a907a8e3a2971e5561d17483fd5b35a not found: ID does not exist" containerID="dba5b7ad175a67c8def72c72279199594a907a8e3a2971e5561d17483fd5b35a" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.221646 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dba5b7ad175a67c8def72c72279199594a907a8e3a2971e5561d17483fd5b35a"} err="failed to get container status \"dba5b7ad175a67c8def72c72279199594a907a8e3a2971e5561d17483fd5b35a\": rpc error: code = NotFound desc = could not find container \"dba5b7ad175a67c8def72c72279199594a907a8e3a2971e5561d17483fd5b35a\": container with ID starting with dba5b7ad175a67c8def72c72279199594a907a8e3a2971e5561d17483fd5b35a not found: ID does not exist" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.221697 4806 scope.go:117] "RemoveContainer" containerID="29c90a7a98e9c0091ec3359bc9822b0da9b54c07f5f7d2a2832d6fba9fb36904" Nov 27 10:41:12 crc kubenswrapper[4806]: E1127 10:41:12.222267 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29c90a7a98e9c0091ec3359bc9822b0da9b54c07f5f7d2a2832d6fba9fb36904\": container with ID starting with 29c90a7a98e9c0091ec3359bc9822b0da9b54c07f5f7d2a2832d6fba9fb36904 not found: ID does not exist" containerID="29c90a7a98e9c0091ec3359bc9822b0da9b54c07f5f7d2a2832d6fba9fb36904" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.222296 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29c90a7a98e9c0091ec3359bc9822b0da9b54c07f5f7d2a2832d6fba9fb36904"} err="failed to get container status \"29c90a7a98e9c0091ec3359bc9822b0da9b54c07f5f7d2a2832d6fba9fb36904\": rpc error: code = NotFound desc = could not find container \"29c90a7a98e9c0091ec3359bc9822b0da9b54c07f5f7d2a2832d6fba9fb36904\": container with ID starting with 29c90a7a98e9c0091ec3359bc9822b0da9b54c07f5f7d2a2832d6fba9fb36904 not found: ID does not exist" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.257186 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jml7x\" (UniqueName: \"kubernetes.io/projected/54c143ac-14ac-45b1-b5a4-0b945167827f-kube-api-access-jml7x\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.257241 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54c143ac-14ac-45b1-b5a4-0b945167827f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.258495 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54c143ac-14ac-45b1-b5a4-0b945167827f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54c143ac-14ac-45b1-b5a4-0b945167827f" (UID: "54c143ac-14ac-45b1-b5a4-0b945167827f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.359197 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54c143ac-14ac-45b1-b5a4-0b945167827f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.380525 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.406368 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.656499 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.176:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 10:41:12 crc kubenswrapper[4806]: I1127 10:41:12.656525 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.176:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.148382 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.187961 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.226969 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.247519 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.270559 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 27 10:41:13 crc kubenswrapper[4806]: E1127 10:41:13.271068 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c143ac-14ac-45b1-b5a4-0b945167827f" containerName="nova-api-log" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.271089 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c143ac-14ac-45b1-b5a4-0b945167827f" containerName="nova-api-log" Nov 27 10:41:13 crc kubenswrapper[4806]: E1127 10:41:13.271110 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c143ac-14ac-45b1-b5a4-0b945167827f" containerName="nova-api-api" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.271117 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c143ac-14ac-45b1-b5a4-0b945167827f" containerName="nova-api-api" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.271373 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="54c143ac-14ac-45b1-b5a4-0b945167827f" containerName="nova-api-api" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.271407 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="54c143ac-14ac-45b1-b5a4-0b945167827f" containerName="nova-api-log" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.272505 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.277895 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.278130 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.278307 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.292332 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.385664 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.385806 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-public-tls-certs\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.385829 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-internal-tls-certs\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.385855 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmwnt\" (UniqueName: \"kubernetes.io/projected/95a6776a-a30b-4ffb-9ba8-51239386cf52-kube-api-access-zmwnt\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.385888 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-config-data\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.385913 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95a6776a-a30b-4ffb-9ba8-51239386cf52-logs\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.488852 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-public-tls-certs\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.489405 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-internal-tls-certs\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.489486 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmwnt\" (UniqueName: \"kubernetes.io/projected/95a6776a-a30b-4ffb-9ba8-51239386cf52-kube-api-access-zmwnt\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.489524 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-config-data\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.489564 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95a6776a-a30b-4ffb-9ba8-51239386cf52-logs\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.489648 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.495521 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-zjnmw"] Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.495878 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95a6776a-a30b-4ffb-9ba8-51239386cf52-logs\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.497151 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.501058 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.501663 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-internal-tls-certs\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.503463 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-config-data\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.505520 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.512655 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-zjnmw"] Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.516995 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmwnt\" (UniqueName: \"kubernetes.io/projected/95a6776a-a30b-4ffb-9ba8-51239386cf52-kube-api-access-zmwnt\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.517661 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-public-tls-certs\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.523839 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.591793 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-scripts\") pod \"nova-cell1-cell-mapping-zjnmw\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.592331 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-zjnmw\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.592484 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-config-data\") pod \"nova-cell1-cell-mapping-zjnmw\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.592632 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhl9s\" (UniqueName: \"kubernetes.io/projected/cce234e3-7d48-4640-8f9d-0e7fac75f77e-kube-api-access-vhl9s\") pod \"nova-cell1-cell-mapping-zjnmw\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.595531 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.695079 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-config-data\") pod \"nova-cell1-cell-mapping-zjnmw\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.697365 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhl9s\" (UniqueName: \"kubernetes.io/projected/cce234e3-7d48-4640-8f9d-0e7fac75f77e-kube-api-access-vhl9s\") pod \"nova-cell1-cell-mapping-zjnmw\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.697497 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-scripts\") pod \"nova-cell1-cell-mapping-zjnmw\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.697760 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-zjnmw\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.703150 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-zjnmw\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.703178 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-config-data\") pod \"nova-cell1-cell-mapping-zjnmw\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.711698 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-scripts\") pod \"nova-cell1-cell-mapping-zjnmw\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.734282 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhl9s\" (UniqueName: \"kubernetes.io/projected/cce234e3-7d48-4640-8f9d-0e7fac75f77e-kube-api-access-vhl9s\") pod \"nova-cell1-cell-mapping-zjnmw\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:13 crc kubenswrapper[4806]: I1127 10:41:13.802625 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:14 crc kubenswrapper[4806]: I1127 10:41:14.165203 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54c143ac-14ac-45b1-b5a4-0b945167827f" path="/var/lib/kubelet/pods/54c143ac-14ac-45b1-b5a4-0b945167827f/volumes" Nov 27 10:41:14 crc kubenswrapper[4806]: I1127 10:41:14.176130 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:41:14 crc kubenswrapper[4806]: I1127 10:41:14.448413 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-zjnmw"] Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.053203 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.144471 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-ceilometer-tls-certs\") pod \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.144581 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-combined-ca-bundle\") pod \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.145562 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-config-data\") pod \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.145948 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-sg-core-conf-yaml\") pod \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.145995 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-log-httpd\") pod \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.146066 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-run-httpd\") pod \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.146178 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-scripts\") pod \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.146318 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgrdt\" (UniqueName: \"kubernetes.io/projected/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-kube-api-access-xgrdt\") pod \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\" (UID: \"94c3d68b-963f-4ba6-b32f-dbbdcaefd377\") " Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.149190 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "94c3d68b-963f-4ba6-b32f-dbbdcaefd377" (UID: "94c3d68b-963f-4ba6-b32f-dbbdcaefd377"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.149870 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "94c3d68b-963f-4ba6-b32f-dbbdcaefd377" (UID: "94c3d68b-963f-4ba6-b32f-dbbdcaefd377"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.169398 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-scripts" (OuterVolumeSpecName: "scripts") pod "94c3d68b-963f-4ba6-b32f-dbbdcaefd377" (UID: "94c3d68b-963f-4ba6-b32f-dbbdcaefd377"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.196660 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-kube-api-access-xgrdt" (OuterVolumeSpecName: "kube-api-access-xgrdt") pod "94c3d68b-963f-4ba6-b32f-dbbdcaefd377" (UID: "94c3d68b-963f-4ba6-b32f-dbbdcaefd377"). InnerVolumeSpecName "kube-api-access-xgrdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.232133 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"95a6776a-a30b-4ffb-9ba8-51239386cf52","Type":"ContainerStarted","Data":"3adac9b02cbcdb0eef0c85e2157059fb3053d326d71431ad6ccfbba88fc80af8"} Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.232191 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"95a6776a-a30b-4ffb-9ba8-51239386cf52","Type":"ContainerStarted","Data":"0964770ea3ce76fc0924ab30ad268c5e3417c872f37cf5b80463a9a024e32cdd"} Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.232204 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"95a6776a-a30b-4ffb-9ba8-51239386cf52","Type":"ContainerStarted","Data":"737bb360fb2f700ce7b0c0edb4d2ccd85ea7eff615ba3d08c947d011233be5b7"} Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.237712 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-zjnmw" event={"ID":"cce234e3-7d48-4640-8f9d-0e7fac75f77e","Type":"ContainerStarted","Data":"c3fe6f748b6629c4e9ba08c580894b58d63e5c15f1fc6e5887449b20e403ce38"} Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.238171 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-zjnmw" event={"ID":"cce234e3-7d48-4640-8f9d-0e7fac75f77e","Type":"ContainerStarted","Data":"ee77ac1333c56b1af1088dc765ceb61afb3069916b5ad5051052a2dcdc9c07a3"} Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.249379 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgrdt\" (UniqueName: \"kubernetes.io/projected/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-kube-api-access-xgrdt\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.249410 4806 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.249419 4806 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.249427 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.252153 4806 generic.go:334] "Generic (PLEG): container finished" podID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerID="0a07048e6b536ac07e6408d46557d83e52d17aae162d8589c4c18fbd247f6b34" exitCode=0 Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.252261 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94c3d68b-963f-4ba6-b32f-dbbdcaefd377","Type":"ContainerDied","Data":"0a07048e6b536ac07e6408d46557d83e52d17aae162d8589c4c18fbd247f6b34"} Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.252299 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94c3d68b-963f-4ba6-b32f-dbbdcaefd377","Type":"ContainerDied","Data":"735c0d4098393563b1f8f5ca747cf7329c22dfa4382e6609b8f1a3679ee2be48"} Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.252340 4806 scope.go:117] "RemoveContainer" containerID="5b571fd767f00ba6ae6b79935a45e658bc009d075a4c472d94a4df45b3648d7a" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.252615 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.258713 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.258693667 podStartE2EDuration="2.258693667s" podCreationTimestamp="2025-11-27 10:41:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:41:15.254638797 +0000 UTC m=+1179.841229561" watchObservedRunningTime="2025-11-27 10:41:15.258693667 +0000 UTC m=+1179.845284431" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.277126 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "94c3d68b-963f-4ba6-b32f-dbbdcaefd377" (UID: "94c3d68b-963f-4ba6-b32f-dbbdcaefd377"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.277520 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "94c3d68b-963f-4ba6-b32f-dbbdcaefd377" (UID: "94c3d68b-963f-4ba6-b32f-dbbdcaefd377"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.288377 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-zjnmw" podStartSLOduration=2.2883555429999998 podStartE2EDuration="2.288355543s" podCreationTimestamp="2025-11-27 10:41:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:41:15.283873001 +0000 UTC m=+1179.870463765" watchObservedRunningTime="2025-11-27 10:41:15.288355543 +0000 UTC m=+1179.874946307" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.310815 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94c3d68b-963f-4ba6-b32f-dbbdcaefd377" (UID: "94c3d68b-963f-4ba6-b32f-dbbdcaefd377"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.312065 4806 scope.go:117] "RemoveContainer" containerID="1511d764c4cd9652da460a249acec5c3439d3acbb23725f75f84b121beee3b3b" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.331096 4806 scope.go:117] "RemoveContainer" containerID="0a07048e6b536ac07e6408d46557d83e52d17aae162d8589c4c18fbd247f6b34" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.345974 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-config-data" (OuterVolumeSpecName: "config-data") pod "94c3d68b-963f-4ba6-b32f-dbbdcaefd377" (UID: "94c3d68b-963f-4ba6-b32f-dbbdcaefd377"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.351042 4806 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.351090 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.351103 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.351113 4806 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94c3d68b-963f-4ba6-b32f-dbbdcaefd377-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.352689 4806 scope.go:117] "RemoveContainer" containerID="0c495272c04e41dccc4c83ccc21e0d26e6d3c18550936024a641225651d6ce0c" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.396551 4806 scope.go:117] "RemoveContainer" containerID="5b571fd767f00ba6ae6b79935a45e658bc009d075a4c472d94a4df45b3648d7a" Nov 27 10:41:15 crc kubenswrapper[4806]: E1127 10:41:15.397112 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b571fd767f00ba6ae6b79935a45e658bc009d075a4c472d94a4df45b3648d7a\": container with ID starting with 5b571fd767f00ba6ae6b79935a45e658bc009d075a4c472d94a4df45b3648d7a not found: ID does not exist" containerID="5b571fd767f00ba6ae6b79935a45e658bc009d075a4c472d94a4df45b3648d7a" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.397156 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b571fd767f00ba6ae6b79935a45e658bc009d075a4c472d94a4df45b3648d7a"} err="failed to get container status \"5b571fd767f00ba6ae6b79935a45e658bc009d075a4c472d94a4df45b3648d7a\": rpc error: code = NotFound desc = could not find container \"5b571fd767f00ba6ae6b79935a45e658bc009d075a4c472d94a4df45b3648d7a\": container with ID starting with 5b571fd767f00ba6ae6b79935a45e658bc009d075a4c472d94a4df45b3648d7a not found: ID does not exist" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.397181 4806 scope.go:117] "RemoveContainer" containerID="1511d764c4cd9652da460a249acec5c3439d3acbb23725f75f84b121beee3b3b" Nov 27 10:41:15 crc kubenswrapper[4806]: E1127 10:41:15.397711 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1511d764c4cd9652da460a249acec5c3439d3acbb23725f75f84b121beee3b3b\": container with ID starting with 1511d764c4cd9652da460a249acec5c3439d3acbb23725f75f84b121beee3b3b not found: ID does not exist" containerID="1511d764c4cd9652da460a249acec5c3439d3acbb23725f75f84b121beee3b3b" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.397740 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1511d764c4cd9652da460a249acec5c3439d3acbb23725f75f84b121beee3b3b"} err="failed to get container status \"1511d764c4cd9652da460a249acec5c3439d3acbb23725f75f84b121beee3b3b\": rpc error: code = NotFound desc = could not find container \"1511d764c4cd9652da460a249acec5c3439d3acbb23725f75f84b121beee3b3b\": container with ID starting with 1511d764c4cd9652da460a249acec5c3439d3acbb23725f75f84b121beee3b3b not found: ID does not exist" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.397755 4806 scope.go:117] "RemoveContainer" containerID="0a07048e6b536ac07e6408d46557d83e52d17aae162d8589c4c18fbd247f6b34" Nov 27 10:41:15 crc kubenswrapper[4806]: E1127 10:41:15.398413 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a07048e6b536ac07e6408d46557d83e52d17aae162d8589c4c18fbd247f6b34\": container with ID starting with 0a07048e6b536ac07e6408d46557d83e52d17aae162d8589c4c18fbd247f6b34 not found: ID does not exist" containerID="0a07048e6b536ac07e6408d46557d83e52d17aae162d8589c4c18fbd247f6b34" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.398453 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a07048e6b536ac07e6408d46557d83e52d17aae162d8589c4c18fbd247f6b34"} err="failed to get container status \"0a07048e6b536ac07e6408d46557d83e52d17aae162d8589c4c18fbd247f6b34\": rpc error: code = NotFound desc = could not find container \"0a07048e6b536ac07e6408d46557d83e52d17aae162d8589c4c18fbd247f6b34\": container with ID starting with 0a07048e6b536ac07e6408d46557d83e52d17aae162d8589c4c18fbd247f6b34 not found: ID does not exist" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.398468 4806 scope.go:117] "RemoveContainer" containerID="0c495272c04e41dccc4c83ccc21e0d26e6d3c18550936024a641225651d6ce0c" Nov 27 10:41:15 crc kubenswrapper[4806]: E1127 10:41:15.398947 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c495272c04e41dccc4c83ccc21e0d26e6d3c18550936024a641225651d6ce0c\": container with ID starting with 0c495272c04e41dccc4c83ccc21e0d26e6d3c18550936024a641225651d6ce0c not found: ID does not exist" containerID="0c495272c04e41dccc4c83ccc21e0d26e6d3c18550936024a641225651d6ce0c" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.398996 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c495272c04e41dccc4c83ccc21e0d26e6d3c18550936024a641225651d6ce0c"} err="failed to get container status \"0c495272c04e41dccc4c83ccc21e0d26e6d3c18550936024a641225651d6ce0c\": rpc error: code = NotFound desc = could not find container \"0c495272c04e41dccc4c83ccc21e0d26e6d3c18550936024a641225651d6ce0c\": container with ID starting with 0c495272c04e41dccc4c83ccc21e0d26e6d3c18550936024a641225651d6ce0c not found: ID does not exist" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.587166 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.601743 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.629214 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:41:15 crc kubenswrapper[4806]: E1127 10:41:15.629732 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="ceilometer-central-agent" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.629755 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="ceilometer-central-agent" Nov 27 10:41:15 crc kubenswrapper[4806]: E1127 10:41:15.629769 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="sg-core" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.629775 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="sg-core" Nov 27 10:41:15 crc kubenswrapper[4806]: E1127 10:41:15.629790 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="proxy-httpd" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.629798 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="proxy-httpd" Nov 27 10:41:15 crc kubenswrapper[4806]: E1127 10:41:15.629808 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="ceilometer-notification-agent" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.629814 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="ceilometer-notification-agent" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.630012 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="ceilometer-notification-agent" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.630026 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="proxy-httpd" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.630046 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="ceilometer-central-agent" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.630054 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" containerName="sg-core" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.646240 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.649570 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.658893 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.659170 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.665426 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.677586 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvgrc\" (UniqueName: \"kubernetes.io/projected/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-kube-api-access-dvgrc\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.677823 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.677983 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-scripts\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.678277 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.678478 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-config-data\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.678737 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.678863 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-run-httpd\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.678996 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-log-httpd\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.741068 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.782869 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.782957 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-config-data\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.783037 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.783066 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-run-httpd\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.783103 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-log-httpd\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.783182 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvgrc\" (UniqueName: \"kubernetes.io/projected/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-kube-api-access-dvgrc\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.783212 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.786386 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-scripts\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.788579 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-log-httpd\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.800513 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-run-httpd\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.801293 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-config-data\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.812993 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.813097 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.813629 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.813697 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-scripts\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.821089 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvgrc\" (UniqueName: \"kubernetes.io/projected/1ebc7600-dc80-48b0-ab6b-01ba5d056c2b-kube-api-access-dvgrc\") pod \"ceilometer-0\" (UID: \"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b\") " pod="openstack/ceilometer-0" Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.838544 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-tdl4c"] Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.838846 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" podUID="d5f6cb41-571f-44d9-b269-33d80225af04" containerName="dnsmasq-dns" containerID="cri-o://6c26baad2629ed7504be46a5ea57bf24f86a5a9c89141fd7e88188558240eea9" gracePeriod=10 Nov 27 10:41:15 crc kubenswrapper[4806]: I1127 10:41:15.974759 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.148990 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94c3d68b-963f-4ba6-b32f-dbbdcaefd377" path="/var/lib/kubelet/pods/94c3d68b-963f-4ba6-b32f-dbbdcaefd377/volumes" Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.274908 4806 generic.go:334] "Generic (PLEG): container finished" podID="d5f6cb41-571f-44d9-b269-33d80225af04" containerID="6c26baad2629ed7504be46a5ea57bf24f86a5a9c89141fd7e88188558240eea9" exitCode=0 Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.275397 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" event={"ID":"d5f6cb41-571f-44d9-b269-33d80225af04","Type":"ContainerDied","Data":"6c26baad2629ed7504be46a5ea57bf24f86a5a9c89141fd7e88188558240eea9"} Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.350577 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.407805 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6k5vf\" (UniqueName: \"kubernetes.io/projected/d5f6cb41-571f-44d9-b269-33d80225af04-kube-api-access-6k5vf\") pod \"d5f6cb41-571f-44d9-b269-33d80225af04\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.407895 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-config\") pod \"d5f6cb41-571f-44d9-b269-33d80225af04\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.407948 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-ovsdbserver-sb\") pod \"d5f6cb41-571f-44d9-b269-33d80225af04\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.407993 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-ovsdbserver-nb\") pod \"d5f6cb41-571f-44d9-b269-33d80225af04\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.408026 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-dns-svc\") pod \"d5f6cb41-571f-44d9-b269-33d80225af04\" (UID: \"d5f6cb41-571f-44d9-b269-33d80225af04\") " Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.419691 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5f6cb41-571f-44d9-b269-33d80225af04-kube-api-access-6k5vf" (OuterVolumeSpecName: "kube-api-access-6k5vf") pod "d5f6cb41-571f-44d9-b269-33d80225af04" (UID: "d5f6cb41-571f-44d9-b269-33d80225af04"). InnerVolumeSpecName "kube-api-access-6k5vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.490263 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-config" (OuterVolumeSpecName: "config") pod "d5f6cb41-571f-44d9-b269-33d80225af04" (UID: "d5f6cb41-571f-44d9-b269-33d80225af04"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.504591 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d5f6cb41-571f-44d9-b269-33d80225af04" (UID: "d5f6cb41-571f-44d9-b269-33d80225af04"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.506358 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d5f6cb41-571f-44d9-b269-33d80225af04" (UID: "d5f6cb41-571f-44d9-b269-33d80225af04"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.511523 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6k5vf\" (UniqueName: \"kubernetes.io/projected/d5f6cb41-571f-44d9-b269-33d80225af04-kube-api-access-6k5vf\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.511561 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.511575 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.511586 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.529024 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d5f6cb41-571f-44d9-b269-33d80225af04" (UID: "d5f6cb41-571f-44d9-b269-33d80225af04"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.572001 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 27 10:41:16 crc kubenswrapper[4806]: W1127 10:41:16.577063 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ebc7600_dc80_48b0_ab6b_01ba5d056c2b.slice/crio-5daca968c29785723d754c39b16b912c4158cdd30a544b0ab2e1fb2659389069 WatchSource:0}: Error finding container 5daca968c29785723d754c39b16b912c4158cdd30a544b0ab2e1fb2659389069: Status 404 returned error can't find the container with id 5daca968c29785723d754c39b16b912c4158cdd30a544b0ab2e1fb2659389069 Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.579998 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 10:41:16 crc kubenswrapper[4806]: I1127 10:41:16.614083 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f6cb41-571f-44d9-b269-33d80225af04-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:17 crc kubenswrapper[4806]: I1127 10:41:17.300252 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" Nov 27 10:41:17 crc kubenswrapper[4806]: I1127 10:41:17.300473 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-tdl4c" event={"ID":"d5f6cb41-571f-44d9-b269-33d80225af04","Type":"ContainerDied","Data":"27c067be34ba9a361855eca26c9478d894489ac7b0fc74fa635b05d0e0e04a8c"} Nov 27 10:41:17 crc kubenswrapper[4806]: I1127 10:41:17.300533 4806 scope.go:117] "RemoveContainer" containerID="6c26baad2629ed7504be46a5ea57bf24f86a5a9c89141fd7e88188558240eea9" Nov 27 10:41:17 crc kubenswrapper[4806]: I1127 10:41:17.309116 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b","Type":"ContainerStarted","Data":"5daca968c29785723d754c39b16b912c4158cdd30a544b0ab2e1fb2659389069"} Nov 27 10:41:17 crc kubenswrapper[4806]: I1127 10:41:17.366305 4806 scope.go:117] "RemoveContainer" containerID="ac0b3b755c9861d50dee1dcfb9431823ffb45a93a2d9881ab7fc7532d4cb3463" Nov 27 10:41:17 crc kubenswrapper[4806]: I1127 10:41:17.384493 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-tdl4c"] Nov 27 10:41:17 crc kubenswrapper[4806]: I1127 10:41:17.396881 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-tdl4c"] Nov 27 10:41:18 crc kubenswrapper[4806]: I1127 10:41:18.137939 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5f6cb41-571f-44d9-b269-33d80225af04" path="/var/lib/kubelet/pods/d5f6cb41-571f-44d9-b269-33d80225af04/volumes" Nov 27 10:41:18 crc kubenswrapper[4806]: I1127 10:41:18.328543 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b","Type":"ContainerStarted","Data":"49a781aa155f47c5077e8acff70bc5a9efc37837d7bcf762dc8cf9caec9a8c1d"} Nov 27 10:41:19 crc kubenswrapper[4806]: I1127 10:41:19.343789 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b","Type":"ContainerStarted","Data":"188a39a07bf7224a7a983ce09b34495cc969211b5e1910cd2b9b311012cd028a"} Nov 27 10:41:19 crc kubenswrapper[4806]: I1127 10:41:19.344555 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b","Type":"ContainerStarted","Data":"d6a0542472ab6d22db6ba626e47a1e446e85ffa4abd86dd86a7a8e4719c86cee"} Nov 27 10:41:21 crc kubenswrapper[4806]: E1127 10:41:21.257818 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce234e3_7d48_4640_8f9d_0e7fac75f77e.slice/crio-c3fe6f748b6629c4e9ba08c580894b58d63e5c15f1fc6e5887449b20e403ce38.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce234e3_7d48_4640_8f9d_0e7fac75f77e.slice/crio-conmon-c3fe6f748b6629c4e9ba08c580894b58d63e5c15f1fc6e5887449b20e403ce38.scope\": RecentStats: unable to find data in memory cache]" Nov 27 10:41:21 crc kubenswrapper[4806]: I1127 10:41:21.392283 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ebc7600-dc80-48b0-ab6b-01ba5d056c2b","Type":"ContainerStarted","Data":"734ba8ef5e7bb72dc99b3fadda04043140e9a62b24e39bf7c8fa74227cabf240"} Nov 27 10:41:21 crc kubenswrapper[4806]: I1127 10:41:21.393585 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 27 10:41:21 crc kubenswrapper[4806]: I1127 10:41:21.396052 4806 generic.go:334] "Generic (PLEG): container finished" podID="cce234e3-7d48-4640-8f9d-0e7fac75f77e" containerID="c3fe6f748b6629c4e9ba08c580894b58d63e5c15f1fc6e5887449b20e403ce38" exitCode=0 Nov 27 10:41:21 crc kubenswrapper[4806]: I1127 10:41:21.396122 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-zjnmw" event={"ID":"cce234e3-7d48-4640-8f9d-0e7fac75f77e","Type":"ContainerDied","Data":"c3fe6f748b6629c4e9ba08c580894b58d63e5c15f1fc6e5887449b20e403ce38"} Nov 27 10:41:21 crc kubenswrapper[4806]: I1127 10:41:21.425109 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.408382754 podStartE2EDuration="6.42507412s" podCreationTimestamp="2025-11-27 10:41:15 +0000 UTC" firstStartedPulling="2025-11-27 10:41:16.579690037 +0000 UTC m=+1181.166280801" lastFinishedPulling="2025-11-27 10:41:20.596381403 +0000 UTC m=+1185.182972167" observedRunningTime="2025-11-27 10:41:21.422745738 +0000 UTC m=+1186.009336502" watchObservedRunningTime="2025-11-27 10:41:21.42507412 +0000 UTC m=+1186.011664884" Nov 27 10:41:21 crc kubenswrapper[4806]: I1127 10:41:21.610947 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 27 10:41:21 crc kubenswrapper[4806]: I1127 10:41:21.613096 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 27 10:41:21 crc kubenswrapper[4806]: I1127 10:41:21.626414 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 27 10:41:22 crc kubenswrapper[4806]: I1127 10:41:22.416608 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 27 10:41:22 crc kubenswrapper[4806]: I1127 10:41:22.838394 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:22 crc kubenswrapper[4806]: I1127 10:41:22.999055 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-scripts\") pod \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " Nov 27 10:41:22 crc kubenswrapper[4806]: I1127 10:41:22.999210 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-combined-ca-bundle\") pod \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " Nov 27 10:41:22 crc kubenswrapper[4806]: I1127 10:41:22.999385 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhl9s\" (UniqueName: \"kubernetes.io/projected/cce234e3-7d48-4640-8f9d-0e7fac75f77e-kube-api-access-vhl9s\") pod \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " Nov 27 10:41:22 crc kubenswrapper[4806]: I1127 10:41:22.999428 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-config-data\") pod \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\" (UID: \"cce234e3-7d48-4640-8f9d-0e7fac75f77e\") " Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.004915 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-scripts" (OuterVolumeSpecName: "scripts") pod "cce234e3-7d48-4640-8f9d-0e7fac75f77e" (UID: "cce234e3-7d48-4640-8f9d-0e7fac75f77e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.006545 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cce234e3-7d48-4640-8f9d-0e7fac75f77e-kube-api-access-vhl9s" (OuterVolumeSpecName: "kube-api-access-vhl9s") pod "cce234e3-7d48-4640-8f9d-0e7fac75f77e" (UID: "cce234e3-7d48-4640-8f9d-0e7fac75f77e"). InnerVolumeSpecName "kube-api-access-vhl9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.032408 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cce234e3-7d48-4640-8f9d-0e7fac75f77e" (UID: "cce234e3-7d48-4640-8f9d-0e7fac75f77e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.039343 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-config-data" (OuterVolumeSpecName: "config-data") pod "cce234e3-7d48-4640-8f9d-0e7fac75f77e" (UID: "cce234e3-7d48-4640-8f9d-0e7fac75f77e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.101878 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhl9s\" (UniqueName: \"kubernetes.io/projected/cce234e3-7d48-4640-8f9d-0e7fac75f77e-kube-api-access-vhl9s\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.101921 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.101931 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-scripts\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.101940 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cce234e3-7d48-4640-8f9d-0e7fac75f77e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.418991 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-zjnmw" event={"ID":"cce234e3-7d48-4640-8f9d-0e7fac75f77e","Type":"ContainerDied","Data":"ee77ac1333c56b1af1088dc765ceb61afb3069916b5ad5051052a2dcdc9c07a3"} Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.419040 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee77ac1333c56b1af1088dc765ceb61afb3069916b5ad5051052a2dcdc9c07a3" Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.419089 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-zjnmw" Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.597339 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.597450 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.655154 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.665519 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.666055 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8dec9511-dd0f-4f87-922a-4959333ec7fc" containerName="nova-scheduler-scheduler" containerID="cri-o://2777010d2ceb747b0fccc30eda6c208196fdba9745e3774019ab2abed1cbb058" gracePeriod=30 Nov 27 10:41:23 crc kubenswrapper[4806]: I1127 10:41:23.737039 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:41:24 crc kubenswrapper[4806]: I1127 10:41:24.428015 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="95a6776a-a30b-4ffb-9ba8-51239386cf52" containerName="nova-api-log" containerID="cri-o://0964770ea3ce76fc0924ab30ad268c5e3417c872f37cf5b80463a9a024e32cdd" gracePeriod=30 Nov 27 10:41:24 crc kubenswrapper[4806]: I1127 10:41:24.428065 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="95a6776a-a30b-4ffb-9ba8-51239386cf52" containerName="nova-api-api" containerID="cri-o://3adac9b02cbcdb0eef0c85e2157059fb3053d326d71431ad6ccfbba88fc80af8" gracePeriod=30 Nov 27 10:41:24 crc kubenswrapper[4806]: I1127 10:41:24.436384 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="95a6776a-a30b-4ffb-9ba8-51239386cf52" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.179:8774/\": EOF" Nov 27 10:41:24 crc kubenswrapper[4806]: I1127 10:41:24.436612 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="95a6776a-a30b-4ffb-9ba8-51239386cf52" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.179:8774/\": EOF" Nov 27 10:41:25 crc kubenswrapper[4806]: I1127 10:41:25.441176 4806 generic.go:334] "Generic (PLEG): container finished" podID="95a6776a-a30b-4ffb-9ba8-51239386cf52" containerID="0964770ea3ce76fc0924ab30ad268c5e3417c872f37cf5b80463a9a024e32cdd" exitCode=143 Nov 27 10:41:25 crc kubenswrapper[4806]: I1127 10:41:25.441934 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerName="nova-metadata-log" containerID="cri-o://0606ae3db73a2db520cbdf66a9662e26d8384e49ed46a484add7af18dae70a47" gracePeriod=30 Nov 27 10:41:25 crc kubenswrapper[4806]: I1127 10:41:25.442280 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"95a6776a-a30b-4ffb-9ba8-51239386cf52","Type":"ContainerDied","Data":"0964770ea3ce76fc0924ab30ad268c5e3417c872f37cf5b80463a9a024e32cdd"} Nov 27 10:41:25 crc kubenswrapper[4806]: I1127 10:41:25.443327 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerName="nova-metadata-metadata" containerID="cri-o://1cdc33f7b797971f6e784bcc81bf09e3cffba7ae86c819edc5255c4ce55ae7d7" gracePeriod=30 Nov 27 10:41:26 crc kubenswrapper[4806]: I1127 10:41:26.455940 4806 generic.go:334] "Generic (PLEG): container finished" podID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerID="0606ae3db73a2db520cbdf66a9662e26d8384e49ed46a484add7af18dae70a47" exitCode=143 Nov 27 10:41:26 crc kubenswrapper[4806]: I1127 10:41:26.456459 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f339550-0757-4afd-b60b-3eacf7d20b4f","Type":"ContainerDied","Data":"0606ae3db73a2db520cbdf66a9662e26d8384e49ed46a484add7af18dae70a47"} Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.255805 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.270588 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dec9511-dd0f-4f87-922a-4959333ec7fc-config-data\") pod \"8dec9511-dd0f-4f87-922a-4959333ec7fc\" (UID: \"8dec9511-dd0f-4f87-922a-4959333ec7fc\") " Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.270656 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dec9511-dd0f-4f87-922a-4959333ec7fc-combined-ca-bundle\") pod \"8dec9511-dd0f-4f87-922a-4959333ec7fc\" (UID: \"8dec9511-dd0f-4f87-922a-4959333ec7fc\") " Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.270771 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9bvm\" (UniqueName: \"kubernetes.io/projected/8dec9511-dd0f-4f87-922a-4959333ec7fc-kube-api-access-z9bvm\") pod \"8dec9511-dd0f-4f87-922a-4959333ec7fc\" (UID: \"8dec9511-dd0f-4f87-922a-4959333ec7fc\") " Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.296577 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dec9511-dd0f-4f87-922a-4959333ec7fc-kube-api-access-z9bvm" (OuterVolumeSpecName: "kube-api-access-z9bvm") pod "8dec9511-dd0f-4f87-922a-4959333ec7fc" (UID: "8dec9511-dd0f-4f87-922a-4959333ec7fc"). InnerVolumeSpecName "kube-api-access-z9bvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.358457 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dec9511-dd0f-4f87-922a-4959333ec7fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8dec9511-dd0f-4f87-922a-4959333ec7fc" (UID: "8dec9511-dd0f-4f87-922a-4959333ec7fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.375108 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dec9511-dd0f-4f87-922a-4959333ec7fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.375163 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9bvm\" (UniqueName: \"kubernetes.io/projected/8dec9511-dd0f-4f87-922a-4959333ec7fc-kube-api-access-z9bvm\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.387429 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dec9511-dd0f-4f87-922a-4959333ec7fc-config-data" (OuterVolumeSpecName: "config-data") pod "8dec9511-dd0f-4f87-922a-4959333ec7fc" (UID: "8dec9511-dd0f-4f87-922a-4959333ec7fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.478390 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dec9511-dd0f-4f87-922a-4959333ec7fc-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.487686 4806 generic.go:334] "Generic (PLEG): container finished" podID="8dec9511-dd0f-4f87-922a-4959333ec7fc" containerID="2777010d2ceb747b0fccc30eda6c208196fdba9745e3774019ab2abed1cbb058" exitCode=0 Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.487754 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8dec9511-dd0f-4f87-922a-4959333ec7fc","Type":"ContainerDied","Data":"2777010d2ceb747b0fccc30eda6c208196fdba9745e3774019ab2abed1cbb058"} Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.487794 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8dec9511-dd0f-4f87-922a-4959333ec7fc","Type":"ContainerDied","Data":"26eb377c96c711ab0a449442efa0101f9158aa199ffc366b5a5825ad17e243c0"} Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.487820 4806 scope.go:117] "RemoveContainer" containerID="2777010d2ceb747b0fccc30eda6c208196fdba9745e3774019ab2abed1cbb058" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.488066 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.521253 4806 scope.go:117] "RemoveContainer" containerID="2777010d2ceb747b0fccc30eda6c208196fdba9745e3774019ab2abed1cbb058" Nov 27 10:41:28 crc kubenswrapper[4806]: E1127 10:41:28.522789 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2777010d2ceb747b0fccc30eda6c208196fdba9745e3774019ab2abed1cbb058\": container with ID starting with 2777010d2ceb747b0fccc30eda6c208196fdba9745e3774019ab2abed1cbb058 not found: ID does not exist" containerID="2777010d2ceb747b0fccc30eda6c208196fdba9745e3774019ab2abed1cbb058" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.522851 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2777010d2ceb747b0fccc30eda6c208196fdba9745e3774019ab2abed1cbb058"} err="failed to get container status \"2777010d2ceb747b0fccc30eda6c208196fdba9745e3774019ab2abed1cbb058\": rpc error: code = NotFound desc = could not find container \"2777010d2ceb747b0fccc30eda6c208196fdba9745e3774019ab2abed1cbb058\": container with ID starting with 2777010d2ceb747b0fccc30eda6c208196fdba9745e3774019ab2abed1cbb058 not found: ID does not exist" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.544364 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.567662 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.577093 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:41:28 crc kubenswrapper[4806]: E1127 10:41:28.581121 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5f6cb41-571f-44d9-b269-33d80225af04" containerName="init" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.581180 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5f6cb41-571f-44d9-b269-33d80225af04" containerName="init" Nov 27 10:41:28 crc kubenswrapper[4806]: E1127 10:41:28.581205 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dec9511-dd0f-4f87-922a-4959333ec7fc" containerName="nova-scheduler-scheduler" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.581214 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dec9511-dd0f-4f87-922a-4959333ec7fc" containerName="nova-scheduler-scheduler" Nov 27 10:41:28 crc kubenswrapper[4806]: E1127 10:41:28.581272 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cce234e3-7d48-4640-8f9d-0e7fac75f77e" containerName="nova-manage" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.581280 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="cce234e3-7d48-4640-8f9d-0e7fac75f77e" containerName="nova-manage" Nov 27 10:41:28 crc kubenswrapper[4806]: E1127 10:41:28.581301 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5f6cb41-571f-44d9-b269-33d80225af04" containerName="dnsmasq-dns" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.581309 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5f6cb41-571f-44d9-b269-33d80225af04" containerName="dnsmasq-dns" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.581519 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dec9511-dd0f-4f87-922a-4959333ec7fc" containerName="nova-scheduler-scheduler" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.581539 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5f6cb41-571f-44d9-b269-33d80225af04" containerName="dnsmasq-dns" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.581549 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="cce234e3-7d48-4640-8f9d-0e7fac75f77e" containerName="nova-manage" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.582612 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.588744 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.599972 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.646709 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.176:8775/\": read tcp 10.217.0.2:39136->10.217.0.176:8775: read: connection reset by peer" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.646708 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.176:8775/\": read tcp 10.217.0.2:39138->10.217.0.176:8775: read: connection reset by peer" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.784994 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69605757-d5b0-4169-9c6e-15f51f282bb3-config-data\") pod \"nova-scheduler-0\" (UID: \"69605757-d5b0-4169-9c6e-15f51f282bb3\") " pod="openstack/nova-scheduler-0" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.787187 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69605757-d5b0-4169-9c6e-15f51f282bb3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"69605757-d5b0-4169-9c6e-15f51f282bb3\") " pod="openstack/nova-scheduler-0" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.787366 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfqzr\" (UniqueName: \"kubernetes.io/projected/69605757-d5b0-4169-9c6e-15f51f282bb3-kube-api-access-cfqzr\") pod \"nova-scheduler-0\" (UID: \"69605757-d5b0-4169-9c6e-15f51f282bb3\") " pod="openstack/nova-scheduler-0" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.888445 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69605757-d5b0-4169-9c6e-15f51f282bb3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"69605757-d5b0-4169-9c6e-15f51f282bb3\") " pod="openstack/nova-scheduler-0" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.888551 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfqzr\" (UniqueName: \"kubernetes.io/projected/69605757-d5b0-4169-9c6e-15f51f282bb3-kube-api-access-cfqzr\") pod \"nova-scheduler-0\" (UID: \"69605757-d5b0-4169-9c6e-15f51f282bb3\") " pod="openstack/nova-scheduler-0" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.888607 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69605757-d5b0-4169-9c6e-15f51f282bb3-config-data\") pod \"nova-scheduler-0\" (UID: \"69605757-d5b0-4169-9c6e-15f51f282bb3\") " pod="openstack/nova-scheduler-0" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.895520 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69605757-d5b0-4169-9c6e-15f51f282bb3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"69605757-d5b0-4169-9c6e-15f51f282bb3\") " pod="openstack/nova-scheduler-0" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.896713 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69605757-d5b0-4169-9c6e-15f51f282bb3-config-data\") pod \"nova-scheduler-0\" (UID: \"69605757-d5b0-4169-9c6e-15f51f282bb3\") " pod="openstack/nova-scheduler-0" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.912982 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfqzr\" (UniqueName: \"kubernetes.io/projected/69605757-d5b0-4169-9c6e-15f51f282bb3-kube-api-access-cfqzr\") pod \"nova-scheduler-0\" (UID: \"69605757-d5b0-4169-9c6e-15f51f282bb3\") " pod="openstack/nova-scheduler-0" Nov 27 10:41:28 crc kubenswrapper[4806]: I1127 10:41:28.933648 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.116018 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.193348 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f339550-0757-4afd-b60b-3eacf7d20b4f-logs\") pod \"9f339550-0757-4afd-b60b-3eacf7d20b4f\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.193446 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-nova-metadata-tls-certs\") pod \"9f339550-0757-4afd-b60b-3eacf7d20b4f\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.193478 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-config-data\") pod \"9f339550-0757-4afd-b60b-3eacf7d20b4f\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.193552 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwm2t\" (UniqueName: \"kubernetes.io/projected/9f339550-0757-4afd-b60b-3eacf7d20b4f-kube-api-access-xwm2t\") pod \"9f339550-0757-4afd-b60b-3eacf7d20b4f\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.193579 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-combined-ca-bundle\") pod \"9f339550-0757-4afd-b60b-3eacf7d20b4f\" (UID: \"9f339550-0757-4afd-b60b-3eacf7d20b4f\") " Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.196738 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f339550-0757-4afd-b60b-3eacf7d20b4f-logs" (OuterVolumeSpecName: "logs") pod "9f339550-0757-4afd-b60b-3eacf7d20b4f" (UID: "9f339550-0757-4afd-b60b-3eacf7d20b4f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.216689 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f339550-0757-4afd-b60b-3eacf7d20b4f-kube-api-access-xwm2t" (OuterVolumeSpecName: "kube-api-access-xwm2t") pod "9f339550-0757-4afd-b60b-3eacf7d20b4f" (UID: "9f339550-0757-4afd-b60b-3eacf7d20b4f"). InnerVolumeSpecName "kube-api-access-xwm2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.230787 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-config-data" (OuterVolumeSpecName: "config-data") pod "9f339550-0757-4afd-b60b-3eacf7d20b4f" (UID: "9f339550-0757-4afd-b60b-3eacf7d20b4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.253926 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f339550-0757-4afd-b60b-3eacf7d20b4f" (UID: "9f339550-0757-4afd-b60b-3eacf7d20b4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.282429 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "9f339550-0757-4afd-b60b-3eacf7d20b4f" (UID: "9f339550-0757-4afd-b60b-3eacf7d20b4f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.296345 4806 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.296382 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.296395 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwm2t\" (UniqueName: \"kubernetes.io/projected/9f339550-0757-4afd-b60b-3eacf7d20b4f-kube-api-access-xwm2t\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.296408 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f339550-0757-4afd-b60b-3eacf7d20b4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.296416 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f339550-0757-4afd-b60b-3eacf7d20b4f-logs\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.392571 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.516422 4806 generic.go:334] "Generic (PLEG): container finished" podID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerID="1cdc33f7b797971f6e784bcc81bf09e3cffba7ae86c819edc5255c4ce55ae7d7" exitCode=0 Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.517027 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f339550-0757-4afd-b60b-3eacf7d20b4f","Type":"ContainerDied","Data":"1cdc33f7b797971f6e784bcc81bf09e3cffba7ae86c819edc5255c4ce55ae7d7"} Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.517125 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.518608 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f339550-0757-4afd-b60b-3eacf7d20b4f","Type":"ContainerDied","Data":"79f4280f26a1b8f84024f89cc0b89ea49019626eec5cd01efb48d2eedcc21c96"} Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.523590 4806 scope.go:117] "RemoveContainer" containerID="1cdc33f7b797971f6e784bcc81bf09e3cffba7ae86c819edc5255c4ce55ae7d7" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.526750 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"69605757-d5b0-4169-9c6e-15f51f282bb3","Type":"ContainerStarted","Data":"d8e78898aba7e10f49588f74270ff5912c1dcf8c9e2a3473308da9a4c7559ea5"} Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.585211 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.604380 4806 scope.go:117] "RemoveContainer" containerID="0606ae3db73a2db520cbdf66a9662e26d8384e49ed46a484add7af18dae70a47" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.604419 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.620402 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:41:29 crc kubenswrapper[4806]: E1127 10:41:29.621481 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerName="nova-metadata-log" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.621507 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerName="nova-metadata-log" Nov 27 10:41:29 crc kubenswrapper[4806]: E1127 10:41:29.621535 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerName="nova-metadata-metadata" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.621541 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerName="nova-metadata-metadata" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.621969 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerName="nova-metadata-metadata" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.622007 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f339550-0757-4afd-b60b-3eacf7d20b4f" containerName="nova-metadata-log" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.623208 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.649784 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.650654 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.653005 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.681529 4806 scope.go:117] "RemoveContainer" containerID="1cdc33f7b797971f6e784bcc81bf09e3cffba7ae86c819edc5255c4ce55ae7d7" Nov 27 10:41:29 crc kubenswrapper[4806]: E1127 10:41:29.682346 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cdc33f7b797971f6e784bcc81bf09e3cffba7ae86c819edc5255c4ce55ae7d7\": container with ID starting with 1cdc33f7b797971f6e784bcc81bf09e3cffba7ae86c819edc5255c4ce55ae7d7 not found: ID does not exist" containerID="1cdc33f7b797971f6e784bcc81bf09e3cffba7ae86c819edc5255c4ce55ae7d7" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.682489 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cdc33f7b797971f6e784bcc81bf09e3cffba7ae86c819edc5255c4ce55ae7d7"} err="failed to get container status \"1cdc33f7b797971f6e784bcc81bf09e3cffba7ae86c819edc5255c4ce55ae7d7\": rpc error: code = NotFound desc = could not find container \"1cdc33f7b797971f6e784bcc81bf09e3cffba7ae86c819edc5255c4ce55ae7d7\": container with ID starting with 1cdc33f7b797971f6e784bcc81bf09e3cffba7ae86c819edc5255c4ce55ae7d7 not found: ID does not exist" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.682598 4806 scope.go:117] "RemoveContainer" containerID="0606ae3db73a2db520cbdf66a9662e26d8384e49ed46a484add7af18dae70a47" Nov 27 10:41:29 crc kubenswrapper[4806]: E1127 10:41:29.686793 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0606ae3db73a2db520cbdf66a9662e26d8384e49ed46a484add7af18dae70a47\": container with ID starting with 0606ae3db73a2db520cbdf66a9662e26d8384e49ed46a484add7af18dae70a47 not found: ID does not exist" containerID="0606ae3db73a2db520cbdf66a9662e26d8384e49ed46a484add7af18dae70a47" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.686857 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0606ae3db73a2db520cbdf66a9662e26d8384e49ed46a484add7af18dae70a47"} err="failed to get container status \"0606ae3db73a2db520cbdf66a9662e26d8384e49ed46a484add7af18dae70a47\": rpc error: code = NotFound desc = could not find container \"0606ae3db73a2db520cbdf66a9662e26d8384e49ed46a484add7af18dae70a47\": container with ID starting with 0606ae3db73a2db520cbdf66a9662e26d8384e49ed46a484add7af18dae70a47 not found: ID does not exist" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.815293 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.815789 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pm5l\" (UniqueName: \"kubernetes.io/projected/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-kube-api-access-9pm5l\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.815826 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-logs\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.815931 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.815986 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-config-data\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.918275 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.918364 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-config-data\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.918432 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.918459 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pm5l\" (UniqueName: \"kubernetes.io/projected/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-kube-api-access-9pm5l\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.918488 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-logs\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.918968 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-logs\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.925744 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-config-data\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.929316 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.936660 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.943336 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pm5l\" (UniqueName: \"kubernetes.io/projected/279060f3-fe83-4add-9a41-d5bf7e4f4e4a-kube-api-access-9pm5l\") pod \"nova-metadata-0\" (UID: \"279060f3-fe83-4add-9a41-d5bf7e4f4e4a\") " pod="openstack/nova-metadata-0" Nov 27 10:41:29 crc kubenswrapper[4806]: I1127 10:41:29.976715 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 27 10:41:30 crc kubenswrapper[4806]: I1127 10:41:30.130161 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dec9511-dd0f-4f87-922a-4959333ec7fc" path="/var/lib/kubelet/pods/8dec9511-dd0f-4f87-922a-4959333ec7fc/volumes" Nov 27 10:41:30 crc kubenswrapper[4806]: I1127 10:41:30.131271 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f339550-0757-4afd-b60b-3eacf7d20b4f" path="/var/lib/kubelet/pods/9f339550-0757-4afd-b60b-3eacf7d20b4f/volumes" Nov 27 10:41:30 crc kubenswrapper[4806]: I1127 10:41:30.476382 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 27 10:41:30 crc kubenswrapper[4806]: I1127 10:41:30.541445 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"279060f3-fe83-4add-9a41-d5bf7e4f4e4a","Type":"ContainerStarted","Data":"bcc6e078f8e17bac82ecd808cf9c030c8ab26d4c59ac71c318ff387047a71a51"} Nov 27 10:41:30 crc kubenswrapper[4806]: I1127 10:41:30.542984 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"69605757-d5b0-4169-9c6e-15f51f282bb3","Type":"ContainerStarted","Data":"8bac8adafbd3609849ad5c56e3595b4e398bf7aed0dd73db115fa191db05d077"} Nov 27 10:41:30 crc kubenswrapper[4806]: I1127 10:41:30.566295 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.566263642 podStartE2EDuration="2.566263642s" podCreationTimestamp="2025-11-27 10:41:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:41:30.561149353 +0000 UTC m=+1195.147740117" watchObservedRunningTime="2025-11-27 10:41:30.566263642 +0000 UTC m=+1195.152854396" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.542209 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.562110 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmwnt\" (UniqueName: \"kubernetes.io/projected/95a6776a-a30b-4ffb-9ba8-51239386cf52-kube-api-access-zmwnt\") pod \"95a6776a-a30b-4ffb-9ba8-51239386cf52\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.562638 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-config-data\") pod \"95a6776a-a30b-4ffb-9ba8-51239386cf52\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.562705 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95a6776a-a30b-4ffb-9ba8-51239386cf52-logs\") pod \"95a6776a-a30b-4ffb-9ba8-51239386cf52\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.562737 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-internal-tls-certs\") pod \"95a6776a-a30b-4ffb-9ba8-51239386cf52\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.562800 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-combined-ca-bundle\") pod \"95a6776a-a30b-4ffb-9ba8-51239386cf52\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.562839 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-public-tls-certs\") pod \"95a6776a-a30b-4ffb-9ba8-51239386cf52\" (UID: \"95a6776a-a30b-4ffb-9ba8-51239386cf52\") " Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.563731 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95a6776a-a30b-4ffb-9ba8-51239386cf52-logs" (OuterVolumeSpecName: "logs") pod "95a6776a-a30b-4ffb-9ba8-51239386cf52" (UID: "95a6776a-a30b-4ffb-9ba8-51239386cf52"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.567202 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"279060f3-fe83-4add-9a41-d5bf7e4f4e4a","Type":"ContainerStarted","Data":"3737fc9e21434ca66b22df9319b827b230f69cc848514741b3ceef158df2e542"} Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.567267 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"279060f3-fe83-4add-9a41-d5bf7e4f4e4a","Type":"ContainerStarted","Data":"90fa993063b1382556056b8d3c10699945c82334cdd951e269b79a5cb00d761f"} Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.571610 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95a6776a-a30b-4ffb-9ba8-51239386cf52-kube-api-access-zmwnt" (OuterVolumeSpecName: "kube-api-access-zmwnt") pod "95a6776a-a30b-4ffb-9ba8-51239386cf52" (UID: "95a6776a-a30b-4ffb-9ba8-51239386cf52"). InnerVolumeSpecName "kube-api-access-zmwnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.573381 4806 generic.go:334] "Generic (PLEG): container finished" podID="95a6776a-a30b-4ffb-9ba8-51239386cf52" containerID="3adac9b02cbcdb0eef0c85e2157059fb3053d326d71431ad6ccfbba88fc80af8" exitCode=0 Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.574032 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.574180 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"95a6776a-a30b-4ffb-9ba8-51239386cf52","Type":"ContainerDied","Data":"3adac9b02cbcdb0eef0c85e2157059fb3053d326d71431ad6ccfbba88fc80af8"} Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.574207 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"95a6776a-a30b-4ffb-9ba8-51239386cf52","Type":"ContainerDied","Data":"737bb360fb2f700ce7b0c0edb4d2ccd85ea7eff615ba3d08c947d011233be5b7"} Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.574224 4806 scope.go:117] "RemoveContainer" containerID="3adac9b02cbcdb0eef0c85e2157059fb3053d326d71431ad6ccfbba88fc80af8" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.654703 4806 scope.go:117] "RemoveContainer" containerID="0964770ea3ce76fc0924ab30ad268c5e3417c872f37cf5b80463a9a024e32cdd" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.682931 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmwnt\" (UniqueName: \"kubernetes.io/projected/95a6776a-a30b-4ffb-9ba8-51239386cf52-kube-api-access-zmwnt\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.682962 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95a6776a-a30b-4ffb-9ba8-51239386cf52-logs\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.687899 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.687882016 podStartE2EDuration="2.687882016s" podCreationTimestamp="2025-11-27 10:41:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:41:31.680271459 +0000 UTC m=+1196.266862223" watchObservedRunningTime="2025-11-27 10:41:31.687882016 +0000 UTC m=+1196.274472780" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.731405 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-config-data" (OuterVolumeSpecName: "config-data") pod "95a6776a-a30b-4ffb-9ba8-51239386cf52" (UID: "95a6776a-a30b-4ffb-9ba8-51239386cf52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.758276 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95a6776a-a30b-4ffb-9ba8-51239386cf52" (UID: "95a6776a-a30b-4ffb-9ba8-51239386cf52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.786609 4806 scope.go:117] "RemoveContainer" containerID="3adac9b02cbcdb0eef0c85e2157059fb3053d326d71431ad6ccfbba88fc80af8" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.801021 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.801051 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:31 crc kubenswrapper[4806]: E1127 10:41:31.801794 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3adac9b02cbcdb0eef0c85e2157059fb3053d326d71431ad6ccfbba88fc80af8\": container with ID starting with 3adac9b02cbcdb0eef0c85e2157059fb3053d326d71431ad6ccfbba88fc80af8 not found: ID does not exist" containerID="3adac9b02cbcdb0eef0c85e2157059fb3053d326d71431ad6ccfbba88fc80af8" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.801837 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3adac9b02cbcdb0eef0c85e2157059fb3053d326d71431ad6ccfbba88fc80af8"} err="failed to get container status \"3adac9b02cbcdb0eef0c85e2157059fb3053d326d71431ad6ccfbba88fc80af8\": rpc error: code = NotFound desc = could not find container \"3adac9b02cbcdb0eef0c85e2157059fb3053d326d71431ad6ccfbba88fc80af8\": container with ID starting with 3adac9b02cbcdb0eef0c85e2157059fb3053d326d71431ad6ccfbba88fc80af8 not found: ID does not exist" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.801870 4806 scope.go:117] "RemoveContainer" containerID="0964770ea3ce76fc0924ab30ad268c5e3417c872f37cf5b80463a9a024e32cdd" Nov 27 10:41:31 crc kubenswrapper[4806]: E1127 10:41:31.811714 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0964770ea3ce76fc0924ab30ad268c5e3417c872f37cf5b80463a9a024e32cdd\": container with ID starting with 0964770ea3ce76fc0924ab30ad268c5e3417c872f37cf5b80463a9a024e32cdd not found: ID does not exist" containerID="0964770ea3ce76fc0924ab30ad268c5e3417c872f37cf5b80463a9a024e32cdd" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.811772 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0964770ea3ce76fc0924ab30ad268c5e3417c872f37cf5b80463a9a024e32cdd"} err="failed to get container status \"0964770ea3ce76fc0924ab30ad268c5e3417c872f37cf5b80463a9a024e32cdd\": rpc error: code = NotFound desc = could not find container \"0964770ea3ce76fc0924ab30ad268c5e3417c872f37cf5b80463a9a024e32cdd\": container with ID starting with 0964770ea3ce76fc0924ab30ad268c5e3417c872f37cf5b80463a9a024e32cdd not found: ID does not exist" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.812386 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "95a6776a-a30b-4ffb-9ba8-51239386cf52" (UID: "95a6776a-a30b-4ffb-9ba8-51239386cf52"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.817165 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "95a6776a-a30b-4ffb-9ba8-51239386cf52" (UID: "95a6776a-a30b-4ffb-9ba8-51239386cf52"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.903585 4806 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.903975 4806 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a6776a-a30b-4ffb-9ba8-51239386cf52-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.914779 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.924665 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.963903 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 27 10:41:31 crc kubenswrapper[4806]: E1127 10:41:31.964969 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a6776a-a30b-4ffb-9ba8-51239386cf52" containerName="nova-api-log" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.964997 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a6776a-a30b-4ffb-9ba8-51239386cf52" containerName="nova-api-log" Nov 27 10:41:31 crc kubenswrapper[4806]: E1127 10:41:31.965042 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a6776a-a30b-4ffb-9ba8-51239386cf52" containerName="nova-api-api" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.965051 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a6776a-a30b-4ffb-9ba8-51239386cf52" containerName="nova-api-api" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.965528 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a6776a-a30b-4ffb-9ba8-51239386cf52" containerName="nova-api-api" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.965564 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a6776a-a30b-4ffb-9ba8-51239386cf52" containerName="nova-api-log" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.974091 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.979510 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.981451 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 27 10:41:31 crc kubenswrapper[4806]: I1127 10:41:31.981848 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.015687 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.110992 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-config-data\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.111075 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vltdx\" (UniqueName: \"kubernetes.io/projected/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-kube-api-access-vltdx\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.111148 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-logs\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.111185 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.111213 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-public-tls-certs\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.111260 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.134721 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95a6776a-a30b-4ffb-9ba8-51239386cf52" path="/var/lib/kubelet/pods/95a6776a-a30b-4ffb-9ba8-51239386cf52/volumes" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.213904 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.213970 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-public-tls-certs\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.213990 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.214098 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-config-data\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.214165 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vltdx\" (UniqueName: \"kubernetes.io/projected/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-kube-api-access-vltdx\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.214223 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-logs\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.214727 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-logs\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.222976 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.230418 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.230873 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-config-data\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.235386 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vltdx\" (UniqueName: \"kubernetes.io/projected/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-kube-api-access-vltdx\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.238698 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41-public-tls-certs\") pod \"nova-api-0\" (UID: \"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41\") " pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.333399 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 27 10:41:32 crc kubenswrapper[4806]: I1127 10:41:32.834170 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 27 10:41:32 crc kubenswrapper[4806]: W1127 10:41:32.838299 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode99ee6cc_a2ad_47d0_b19b_b15fc1c48d41.slice/crio-0646bf1da9e27ab3bdf3cd10a4a7b8239239f757192bee930f9917e433ab184a WatchSource:0}: Error finding container 0646bf1da9e27ab3bdf3cd10a4a7b8239239f757192bee930f9917e433ab184a: Status 404 returned error can't find the container with id 0646bf1da9e27ab3bdf3cd10a4a7b8239239f757192bee930f9917e433ab184a Nov 27 10:41:33 crc kubenswrapper[4806]: I1127 10:41:33.598206 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41","Type":"ContainerStarted","Data":"42d1800ef2b9f0cc9a95eef0f8b77e7e633b68573d3f1f7d88b7a1d04577e293"} Nov 27 10:41:33 crc kubenswrapper[4806]: I1127 10:41:33.598668 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41","Type":"ContainerStarted","Data":"83a3dd002e3363cd441f4c2c4837d4d6369e9a62bf5affe93d0d094c42c52774"} Nov 27 10:41:33 crc kubenswrapper[4806]: I1127 10:41:33.598685 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41","Type":"ContainerStarted","Data":"0646bf1da9e27ab3bdf3cd10a4a7b8239239f757192bee930f9917e433ab184a"} Nov 27 10:41:33 crc kubenswrapper[4806]: I1127 10:41:33.625340 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.625310308 podStartE2EDuration="2.625310308s" podCreationTimestamp="2025-11-27 10:41:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:41:33.620868687 +0000 UTC m=+1198.207459491" watchObservedRunningTime="2025-11-27 10:41:33.625310308 +0000 UTC m=+1198.211901062" Nov 27 10:41:33 crc kubenswrapper[4806]: I1127 10:41:33.934109 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 27 10:41:34 crc kubenswrapper[4806]: I1127 10:41:34.977131 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 10:41:34 crc kubenswrapper[4806]: I1127 10:41:34.977565 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 27 10:41:38 crc kubenswrapper[4806]: I1127 10:41:38.934630 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 27 10:41:38 crc kubenswrapper[4806]: I1127 10:41:38.962646 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 27 10:41:39 crc kubenswrapper[4806]: I1127 10:41:39.873595 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 27 10:41:39 crc kubenswrapper[4806]: I1127 10:41:39.977120 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 27 10:41:39 crc kubenswrapper[4806]: I1127 10:41:39.977188 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 27 10:41:40 crc kubenswrapper[4806]: I1127 10:41:40.985506 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="279060f3-fe83-4add-9a41-d5bf7e4f4e4a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.183:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 10:41:40 crc kubenswrapper[4806]: I1127 10:41:40.985548 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="279060f3-fe83-4add-9a41-d5bf7e4f4e4a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.183:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 10:41:42 crc kubenswrapper[4806]: I1127 10:41:42.335157 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 10:41:42 crc kubenswrapper[4806]: I1127 10:41:42.335557 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 27 10:41:43 crc kubenswrapper[4806]: I1127 10:41:43.351647 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.184:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 10:41:43 crc kubenswrapper[4806]: I1127 10:41:43.351658 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.184:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 27 10:41:45 crc kubenswrapper[4806]: I1127 10:41:45.985480 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 27 10:41:49 crc kubenswrapper[4806]: I1127 10:41:49.983371 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 27 10:41:49 crc kubenswrapper[4806]: I1127 10:41:49.991731 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 27 10:41:49 crc kubenswrapper[4806]: I1127 10:41:49.994401 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 27 10:41:50 crc kubenswrapper[4806]: I1127 10:41:50.790621 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 27 10:41:52 crc kubenswrapper[4806]: I1127 10:41:52.341627 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 27 10:41:52 crc kubenswrapper[4806]: I1127 10:41:52.342169 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 27 10:41:52 crc kubenswrapper[4806]: I1127 10:41:52.342593 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 27 10:41:52 crc kubenswrapper[4806]: I1127 10:41:52.342641 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 27 10:41:52 crc kubenswrapper[4806]: I1127 10:41:52.351431 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 27 10:41:52 crc kubenswrapper[4806]: I1127 10:41:52.356829 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 27 10:42:00 crc kubenswrapper[4806]: I1127 10:42:00.795201 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 10:42:01 crc kubenswrapper[4806]: I1127 10:42:01.837355 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 10:42:06 crc kubenswrapper[4806]: I1127 10:42:06.374900 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="08137128-e0d1-49d2-abf1-1242c1cb58ce" containerName="rabbitmq" containerID="cri-o://fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea" gracePeriod=604795 Nov 27 10:42:07 crc kubenswrapper[4806]: I1127 10:42:07.597038 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="954f26d7-2bdc-45e3-a07b-d9924469f0a1" containerName="rabbitmq" containerID="cri-o://20acdf98b7f0a69423bc1346d4f34b766d96fc603ba847c5f00d8813173ee867" gracePeriod=604795 Nov 27 10:42:11 crc kubenswrapper[4806]: I1127 10:42:11.679344 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="08137128-e0d1-49d2-abf1-1242c1cb58ce" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.94:5671: connect: connection refused" Nov 27 10:42:11 crc kubenswrapper[4806]: I1127 10:42:11.820469 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="954f26d7-2bdc-45e3-a07b-d9924469f0a1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.95:5671: connect: connection refused" Nov 27 10:42:12 crc kubenswrapper[4806]: E1127 10:42:12.659291 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08137128_e0d1_49d2_abf1_1242c1cb58ce.slice/crio-fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea.scope\": RecentStats: unable to find data in memory cache]" Nov 27 10:42:12 crc kubenswrapper[4806]: I1127 10:42:12.980432 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.038114 4806 generic.go:334] "Generic (PLEG): container finished" podID="08137128-e0d1-49d2-abf1-1242c1cb58ce" containerID="fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea" exitCode=0 Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.038172 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"08137128-e0d1-49d2-abf1-1242c1cb58ce","Type":"ContainerDied","Data":"fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea"} Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.038213 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"08137128-e0d1-49d2-abf1-1242c1cb58ce","Type":"ContainerDied","Data":"a39aa1d2e145b761ba1c1c263afbca6fea9593d6c18b6dd24b3047a20a4e09fc"} Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.038250 4806 scope.go:117] "RemoveContainer" containerID="fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.038406 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.101085 4806 scope.go:117] "RemoveContainer" containerID="e7c3736208836a8b527976f788183343ce7a097c822f342b20cd72859c3b83dd" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.121945 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/08137128-e0d1-49d2-abf1-1242c1cb58ce-erlang-cookie-secret\") pod \"08137128-e0d1-49d2-abf1-1242c1cb58ce\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.122068 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/08137128-e0d1-49d2-abf1-1242c1cb58ce-pod-info\") pod \"08137128-e0d1-49d2-abf1-1242c1cb58ce\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.122583 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ncnr\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-kube-api-access-7ncnr\") pod \"08137128-e0d1-49d2-abf1-1242c1cb58ce\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.122637 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-plugins\") pod \"08137128-e0d1-49d2-abf1-1242c1cb58ce\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.124022 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"08137128-e0d1-49d2-abf1-1242c1cb58ce\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.124152 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-tls\") pod \"08137128-e0d1-49d2-abf1-1242c1cb58ce\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.124180 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-erlang-cookie\") pod \"08137128-e0d1-49d2-abf1-1242c1cb58ce\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.124217 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-server-conf\") pod \"08137128-e0d1-49d2-abf1-1242c1cb58ce\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.124253 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-plugins-conf\") pod \"08137128-e0d1-49d2-abf1-1242c1cb58ce\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.124289 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-confd\") pod \"08137128-e0d1-49d2-abf1-1242c1cb58ce\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.124334 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-config-data\") pod \"08137128-e0d1-49d2-abf1-1242c1cb58ce\" (UID: \"08137128-e0d1-49d2-abf1-1242c1cb58ce\") " Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.132248 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "08137128-e0d1-49d2-abf1-1242c1cb58ce" (UID: "08137128-e0d1-49d2-abf1-1242c1cb58ce"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.136986 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "08137128-e0d1-49d2-abf1-1242c1cb58ce" (UID: "08137128-e0d1-49d2-abf1-1242c1cb58ce"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.140850 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "08137128-e0d1-49d2-abf1-1242c1cb58ce" (UID: "08137128-e0d1-49d2-abf1-1242c1cb58ce"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.178118 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08137128-e0d1-49d2-abf1-1242c1cb58ce-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "08137128-e0d1-49d2-abf1-1242c1cb58ce" (UID: "08137128-e0d1-49d2-abf1-1242c1cb58ce"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.197317 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-kube-api-access-7ncnr" (OuterVolumeSpecName: "kube-api-access-7ncnr") pod "08137128-e0d1-49d2-abf1-1242c1cb58ce" (UID: "08137128-e0d1-49d2-abf1-1242c1cb58ce"). InnerVolumeSpecName "kube-api-access-7ncnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.197381 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "08137128-e0d1-49d2-abf1-1242c1cb58ce" (UID: "08137128-e0d1-49d2-abf1-1242c1cb58ce"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.197629 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "08137128-e0d1-49d2-abf1-1242c1cb58ce" (UID: "08137128-e0d1-49d2-abf1-1242c1cb58ce"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.198082 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/08137128-e0d1-49d2-abf1-1242c1cb58ce-pod-info" (OuterVolumeSpecName: "pod-info") pod "08137128-e0d1-49d2-abf1-1242c1cb58ce" (UID: "08137128-e0d1-49d2-abf1-1242c1cb58ce"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.232566 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.232612 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.232623 4806 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.232635 4806 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/08137128-e0d1-49d2-abf1-1242c1cb58ce-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.232643 4806 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/08137128-e0d1-49d2-abf1-1242c1cb58ce-pod-info\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.232651 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ncnr\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-kube-api-access-7ncnr\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.232659 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.232681 4806 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.273306 4806 scope.go:117] "RemoveContainer" containerID="fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea" Nov 27 10:42:13 crc kubenswrapper[4806]: E1127 10:42:13.275461 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea\": container with ID starting with fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea not found: ID does not exist" containerID="fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.275496 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea"} err="failed to get container status \"fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea\": rpc error: code = NotFound desc = could not find container \"fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea\": container with ID starting with fceba88bf732479c05f645ac880b43037aefc51015225b02dfd5b0479a8622ea not found: ID does not exist" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.275523 4806 scope.go:117] "RemoveContainer" containerID="e7c3736208836a8b527976f788183343ce7a097c822f342b20cd72859c3b83dd" Nov 27 10:42:13 crc kubenswrapper[4806]: E1127 10:42:13.282005 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7c3736208836a8b527976f788183343ce7a097c822f342b20cd72859c3b83dd\": container with ID starting with e7c3736208836a8b527976f788183343ce7a097c822f342b20cd72859c3b83dd not found: ID does not exist" containerID="e7c3736208836a8b527976f788183343ce7a097c822f342b20cd72859c3b83dd" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.282054 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7c3736208836a8b527976f788183343ce7a097c822f342b20cd72859c3b83dd"} err="failed to get container status \"e7c3736208836a8b527976f788183343ce7a097c822f342b20cd72859c3b83dd\": rpc error: code = NotFound desc = could not find container \"e7c3736208836a8b527976f788183343ce7a097c822f342b20cd72859c3b83dd\": container with ID starting with e7c3736208836a8b527976f788183343ce7a097c822f342b20cd72859c3b83dd not found: ID does not exist" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.282181 4806 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.291560 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-config-data" (OuterVolumeSpecName: "config-data") pod "08137128-e0d1-49d2-abf1-1242c1cb58ce" (UID: "08137128-e0d1-49d2-abf1-1242c1cb58ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.306157 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-server-conf" (OuterVolumeSpecName: "server-conf") pod "08137128-e0d1-49d2-abf1-1242c1cb58ce" (UID: "08137128-e0d1-49d2-abf1-1242c1cb58ce"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.336557 4806 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.336618 4806 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-server-conf\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.336644 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08137128-e0d1-49d2-abf1-1242c1cb58ce-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.367789 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "08137128-e0d1-49d2-abf1-1242c1cb58ce" (UID: "08137128-e0d1-49d2-abf1-1242c1cb58ce"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.439090 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/08137128-e0d1-49d2-abf1-1242c1cb58ce-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.696402 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.706720 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.733158 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 10:42:13 crc kubenswrapper[4806]: E1127 10:42:13.733781 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08137128-e0d1-49d2-abf1-1242c1cb58ce" containerName="rabbitmq" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.733814 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="08137128-e0d1-49d2-abf1-1242c1cb58ce" containerName="rabbitmq" Nov 27 10:42:13 crc kubenswrapper[4806]: E1127 10:42:13.733836 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08137128-e0d1-49d2-abf1-1242c1cb58ce" containerName="setup-container" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.733845 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="08137128-e0d1-49d2-abf1-1242c1cb58ce" containerName="setup-container" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.734128 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="08137128-e0d1-49d2-abf1-1242c1cb58ce" containerName="rabbitmq" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.735690 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.738556 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.739075 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-rhfhd" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.740010 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.740769 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.741038 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.741487 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.746081 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.773980 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.850545 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5052338d-2912-432a-a022-aa169c1d0930-config-data\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.850604 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5052338d-2912-432a-a022-aa169c1d0930-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.850956 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5052338d-2912-432a-a022-aa169c1d0930-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.851029 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.851058 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5052338d-2912-432a-a022-aa169c1d0930-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.851132 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5052338d-2912-432a-a022-aa169c1d0930-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.851165 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbljn\" (UniqueName: \"kubernetes.io/projected/5052338d-2912-432a-a022-aa169c1d0930-kube-api-access-zbljn\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.851214 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5052338d-2912-432a-a022-aa169c1d0930-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.851276 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5052338d-2912-432a-a022-aa169c1d0930-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.851323 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5052338d-2912-432a-a022-aa169c1d0930-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.851352 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5052338d-2912-432a-a022-aa169c1d0930-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.958799 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5052338d-2912-432a-a022-aa169c1d0930-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.958859 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbljn\" (UniqueName: \"kubernetes.io/projected/5052338d-2912-432a-a022-aa169c1d0930-kube-api-access-zbljn\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.958909 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5052338d-2912-432a-a022-aa169c1d0930-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.958936 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5052338d-2912-432a-a022-aa169c1d0930-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.958966 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5052338d-2912-432a-a022-aa169c1d0930-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.958989 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5052338d-2912-432a-a022-aa169c1d0930-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.959068 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5052338d-2912-432a-a022-aa169c1d0930-config-data\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.959087 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5052338d-2912-432a-a022-aa169c1d0930-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.959103 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5052338d-2912-432a-a022-aa169c1d0930-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.959129 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.959153 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5052338d-2912-432a-a022-aa169c1d0930-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.959883 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5052338d-2912-432a-a022-aa169c1d0930-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.960059 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5052338d-2912-432a-a022-aa169c1d0930-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.960088 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5052338d-2912-432a-a022-aa169c1d0930-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.960340 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.964810 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5052338d-2912-432a-a022-aa169c1d0930-config-data\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.969490 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5052338d-2912-432a-a022-aa169c1d0930-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.974397 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5052338d-2912-432a-a022-aa169c1d0930-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.975137 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5052338d-2912-432a-a022-aa169c1d0930-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.975980 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5052338d-2912-432a-a022-aa169c1d0930-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.979343 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbljn\" (UniqueName: \"kubernetes.io/projected/5052338d-2912-432a-a022-aa169c1d0930-kube-api-access-zbljn\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:13 crc kubenswrapper[4806]: I1127 10:42:13.993749 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5052338d-2912-432a-a022-aa169c1d0930-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.048415 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"5052338d-2912-432a-a022-aa169c1d0930\") " pod="openstack/rabbitmq-server-0" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.100026 4806 generic.go:334] "Generic (PLEG): container finished" podID="954f26d7-2bdc-45e3-a07b-d9924469f0a1" containerID="20acdf98b7f0a69423bc1346d4f34b766d96fc603ba847c5f00d8813173ee867" exitCode=0 Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.100125 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"954f26d7-2bdc-45e3-a07b-d9924469f0a1","Type":"ContainerDied","Data":"20acdf98b7f0a69423bc1346d4f34b766d96fc603ba847c5f00d8813173ee867"} Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.187177 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08137128-e0d1-49d2-abf1-1242c1cb58ce" path="/var/lib/kubelet/pods/08137128-e0d1-49d2-abf1-1242c1cb58ce/volumes" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.364520 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.396321 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.450815 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.450890 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.517780 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnwtp\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-kube-api-access-bnwtp\") pod \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.517968 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/954f26d7-2bdc-45e3-a07b-d9924469f0a1-erlang-cookie-secret\") pod \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.518041 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-server-conf\") pod \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.518135 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/954f26d7-2bdc-45e3-a07b-d9924469f0a1-pod-info\") pod \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.518168 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.518202 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-plugins-conf\") pod \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.518287 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-tls\") pod \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.520710 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "954f26d7-2bdc-45e3-a07b-d9924469f0a1" (UID: "954f26d7-2bdc-45e3-a07b-d9924469f0a1"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.520834 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-config-data\") pod \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.520897 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-erlang-cookie\") pod \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.520936 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-plugins\") pod \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.520960 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-confd\") pod \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\" (UID: \"954f26d7-2bdc-45e3-a07b-d9924469f0a1\") " Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.522895 4806 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.526512 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "954f26d7-2bdc-45e3-a07b-d9924469f0a1" (UID: "954f26d7-2bdc-45e3-a07b-d9924469f0a1"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.526747 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "954f26d7-2bdc-45e3-a07b-d9924469f0a1" (UID: "954f26d7-2bdc-45e3-a07b-d9924469f0a1"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.528039 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/954f26d7-2bdc-45e3-a07b-d9924469f0a1-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "954f26d7-2bdc-45e3-a07b-d9924469f0a1" (UID: "954f26d7-2bdc-45e3-a07b-d9924469f0a1"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.551311 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-kube-api-access-bnwtp" (OuterVolumeSpecName: "kube-api-access-bnwtp") pod "954f26d7-2bdc-45e3-a07b-d9924469f0a1" (UID: "954f26d7-2bdc-45e3-a07b-d9924469f0a1"). InnerVolumeSpecName "kube-api-access-bnwtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.555103 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "954f26d7-2bdc-45e3-a07b-d9924469f0a1" (UID: "954f26d7-2bdc-45e3-a07b-d9924469f0a1"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.555262 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "954f26d7-2bdc-45e3-a07b-d9924469f0a1" (UID: "954f26d7-2bdc-45e3-a07b-d9924469f0a1"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.576526 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/954f26d7-2bdc-45e3-a07b-d9924469f0a1-pod-info" (OuterVolumeSpecName: "pod-info") pod "954f26d7-2bdc-45e3-a07b-d9924469f0a1" (UID: "954f26d7-2bdc-45e3-a07b-d9924469f0a1"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.609417 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-config-data" (OuterVolumeSpecName: "config-data") pod "954f26d7-2bdc-45e3-a07b-d9924469f0a1" (UID: "954f26d7-2bdc-45e3-a07b-d9924469f0a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.625185 4806 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/954f26d7-2bdc-45e3-a07b-d9924469f0a1-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.643095 4806 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/954f26d7-2bdc-45e3-a07b-d9924469f0a1-pod-info\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.643168 4806 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.651022 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.652412 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.652522 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.652590 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.652728 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnwtp\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-kube-api-access-bnwtp\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.653819 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-server-conf" (OuterVolumeSpecName: "server-conf") pod "954f26d7-2bdc-45e3-a07b-d9924469f0a1" (UID: "954f26d7-2bdc-45e3-a07b-d9924469f0a1"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.717459 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "954f26d7-2bdc-45e3-a07b-d9924469f0a1" (UID: "954f26d7-2bdc-45e3-a07b-d9924469f0a1"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.725437 4806 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.754616 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/954f26d7-2bdc-45e3-a07b-d9924469f0a1-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.754673 4806 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/954f26d7-2bdc-45e3-a07b-d9924469f0a1-server-conf\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:14 crc kubenswrapper[4806]: I1127 10:42:14.754683 4806 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.041018 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.133504 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5052338d-2912-432a-a022-aa169c1d0930","Type":"ContainerStarted","Data":"4a0d6573d7e3f1c9bae5b5a66fa0970cdf77fa9033be25e5f21b7523de851c81"} Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.138401 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"954f26d7-2bdc-45e3-a07b-d9924469f0a1","Type":"ContainerDied","Data":"9ec78ae4e8d5273abfdda51f5cdf42fb72508d56b70ca2134e687b9e5b6f156b"} Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.138636 4806 scope.go:117] "RemoveContainer" containerID="20acdf98b7f0a69423bc1346d4f34b766d96fc603ba847c5f00d8813173ee867" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.138456 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.175885 4806 scope.go:117] "RemoveContainer" containerID="96a9494ee6564f52a58e01b832a4f8f8d3d7c39e07840c0adfcef6d9414c5d1b" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.194638 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.211976 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.228907 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 10:42:15 crc kubenswrapper[4806]: E1127 10:42:15.229383 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954f26d7-2bdc-45e3-a07b-d9924469f0a1" containerName="rabbitmq" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.229401 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="954f26d7-2bdc-45e3-a07b-d9924469f0a1" containerName="rabbitmq" Nov 27 10:42:15 crc kubenswrapper[4806]: E1127 10:42:15.229426 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954f26d7-2bdc-45e3-a07b-d9924469f0a1" containerName="setup-container" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.229434 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="954f26d7-2bdc-45e3-a07b-d9924469f0a1" containerName="setup-container" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.230596 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="954f26d7-2bdc-45e3-a07b-d9924469f0a1" containerName="rabbitmq" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.231654 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.235451 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.239882 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.240057 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.240252 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-dl47t" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.240379 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.240523 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.240667 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.250560 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.394437 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.394507 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3cdb0a58-54e1-4d11-8240-721304f548dd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.394536 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cdb0a58-54e1-4d11-8240-721304f548dd-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.394583 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcqd9\" (UniqueName: \"kubernetes.io/projected/3cdb0a58-54e1-4d11-8240-721304f548dd-kube-api-access-dcqd9\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.394601 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3cdb0a58-54e1-4d11-8240-721304f548dd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.394682 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3cdb0a58-54e1-4d11-8240-721304f548dd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.394729 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3cdb0a58-54e1-4d11-8240-721304f548dd-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.394760 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3cdb0a58-54e1-4d11-8240-721304f548dd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.394790 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3cdb0a58-54e1-4d11-8240-721304f548dd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.394858 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3cdb0a58-54e1-4d11-8240-721304f548dd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.394905 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3cdb0a58-54e1-4d11-8240-721304f548dd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.496774 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3cdb0a58-54e1-4d11-8240-721304f548dd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.496841 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cdb0a58-54e1-4d11-8240-721304f548dd-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.496885 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcqd9\" (UniqueName: \"kubernetes.io/projected/3cdb0a58-54e1-4d11-8240-721304f548dd-kube-api-access-dcqd9\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.496906 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3cdb0a58-54e1-4d11-8240-721304f548dd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.496947 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3cdb0a58-54e1-4d11-8240-721304f548dd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.496975 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3cdb0a58-54e1-4d11-8240-721304f548dd-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.496991 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3cdb0a58-54e1-4d11-8240-721304f548dd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.497041 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3cdb0a58-54e1-4d11-8240-721304f548dd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.497089 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3cdb0a58-54e1-4d11-8240-721304f548dd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.497122 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3cdb0a58-54e1-4d11-8240-721304f548dd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.497149 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.497770 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.498425 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3cdb0a58-54e1-4d11-8240-721304f548dd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.498844 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3cdb0a58-54e1-4d11-8240-721304f548dd-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.498917 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3cdb0a58-54e1-4d11-8240-721304f548dd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.499170 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3cdb0a58-54e1-4d11-8240-721304f548dd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.499803 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3cdb0a58-54e1-4d11-8240-721304f548dd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.502370 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3cdb0a58-54e1-4d11-8240-721304f548dd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.502946 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3cdb0a58-54e1-4d11-8240-721304f548dd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.502992 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3cdb0a58-54e1-4d11-8240-721304f548dd-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.504614 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3cdb0a58-54e1-4d11-8240-721304f548dd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.522969 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcqd9\" (UniqueName: \"kubernetes.io/projected/3cdb0a58-54e1-4d11-8240-721304f548dd-kube-api-access-dcqd9\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.532855 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3cdb0a58-54e1-4d11-8240-721304f548dd\") " pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:15 crc kubenswrapper[4806]: I1127 10:42:15.555219 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:16 crc kubenswrapper[4806]: I1127 10:42:16.055103 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 27 10:42:16 crc kubenswrapper[4806]: I1127 10:42:16.136677 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="954f26d7-2bdc-45e3-a07b-d9924469f0a1" path="/var/lib/kubelet/pods/954f26d7-2bdc-45e3-a07b-d9924469f0a1/volumes" Nov 27 10:42:16 crc kubenswrapper[4806]: I1127 10:42:16.170532 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3cdb0a58-54e1-4d11-8240-721304f548dd","Type":"ContainerStarted","Data":"1e2cc07df41148358241b0337e82c3e80a9101e68a3467c11a2457edce9c9c09"} Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.016673 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-t2ptf"] Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.018944 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.026870 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-t2ptf"] Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.030475 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.134914 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-config\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.134998 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.135061 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-dns-svc\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.135099 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hf4w\" (UniqueName: \"kubernetes.io/projected/615d096c-c354-4ba0-ad6c-c1b25e50033a-kube-api-access-6hf4w\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.135138 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.135167 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.237116 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-config\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.237198 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.237303 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-dns-svc\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.237354 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hf4w\" (UniqueName: \"kubernetes.io/projected/615d096c-c354-4ba0-ad6c-c1b25e50033a-kube-api-access-6hf4w\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.237407 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.237449 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.238316 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-config\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.238873 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.239220 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.239435 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-dns-svc\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.239447 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.275707 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hf4w\" (UniqueName: \"kubernetes.io/projected/615d096c-c354-4ba0-ad6c-c1b25e50033a-kube-api-access-6hf4w\") pod \"dnsmasq-dns-578b8d767c-t2ptf\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.376587 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:17 crc kubenswrapper[4806]: I1127 10:42:17.894695 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-t2ptf"] Nov 27 10:42:18 crc kubenswrapper[4806]: I1127 10:42:18.195581 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5052338d-2912-432a-a022-aa169c1d0930","Type":"ContainerStarted","Data":"fcb13c64aaa4a1a540bda08dae7b399d51a62b28720239fe07254f98f97c913d"} Nov 27 10:42:18 crc kubenswrapper[4806]: I1127 10:42:18.199174 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3cdb0a58-54e1-4d11-8240-721304f548dd","Type":"ContainerStarted","Data":"5af461ece93a88fa27e8d6a182e6025d312ee415c761bf0fa53683c5adb6df88"} Nov 27 10:42:18 crc kubenswrapper[4806]: I1127 10:42:18.202539 4806 generic.go:334] "Generic (PLEG): container finished" podID="615d096c-c354-4ba0-ad6c-c1b25e50033a" containerID="fa72edb9e27c6e415909e9806dac76c58bbeed37d5de6af3b1e2133f8d4dbb94" exitCode=0 Nov 27 10:42:18 crc kubenswrapper[4806]: I1127 10:42:18.202577 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" event={"ID":"615d096c-c354-4ba0-ad6c-c1b25e50033a","Type":"ContainerDied","Data":"fa72edb9e27c6e415909e9806dac76c58bbeed37d5de6af3b1e2133f8d4dbb94"} Nov 27 10:42:18 crc kubenswrapper[4806]: I1127 10:42:18.202597 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" event={"ID":"615d096c-c354-4ba0-ad6c-c1b25e50033a","Type":"ContainerStarted","Data":"2926dcde78f8fc9e108b8f319ab30c09bd51a954f119d7597f206dcf7d5d802e"} Nov 27 10:42:19 crc kubenswrapper[4806]: I1127 10:42:19.216657 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" event={"ID":"615d096c-c354-4ba0-ad6c-c1b25e50033a","Type":"ContainerStarted","Data":"d770fcdeac8209d8bce8db8ce69c34bccd8e4205cd4473341b20f4502c7eb3fa"} Nov 27 10:42:19 crc kubenswrapper[4806]: I1127 10:42:19.243071 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" podStartSLOduration=3.243042606 podStartE2EDuration="3.243042606s" podCreationTimestamp="2025-11-27 10:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:42:19.241479343 +0000 UTC m=+1243.828070117" watchObservedRunningTime="2025-11-27 10:42:19.243042606 +0000 UTC m=+1243.829633370" Nov 27 10:42:20 crc kubenswrapper[4806]: I1127 10:42:20.227941 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.378114 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.465443 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-dkkxb"] Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.467053 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" podUID="356b8268-4acb-4ccd-92c5-b085c9fdd884" containerName="dnsmasq-dns" containerID="cri-o://14fde9f23ade1454eddf0c593d445bad7df3fa72e3f58ec0d0de0fc4368be60c" gracePeriod=10 Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.701331 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-667ff9c869-wcdvm"] Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.703643 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.726450 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-667ff9c869-wcdvm"] Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.782677 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-ovsdbserver-sb\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.782739 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-ovsdbserver-nb\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.782919 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-dns-svc\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.783087 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b96j2\" (UniqueName: \"kubernetes.io/projected/c4420782-e4e9-4241-9083-de9ace25d296-kube-api-access-b96j2\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.783350 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-config\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.783455 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-openstack-edpm-ipam\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.885492 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-ovsdbserver-nb\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.885552 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-dns-svc\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.885582 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b96j2\" (UniqueName: \"kubernetes.io/projected/c4420782-e4e9-4241-9083-de9ace25d296-kube-api-access-b96j2\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.885625 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-config\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.885654 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-openstack-edpm-ipam\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.885742 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-ovsdbserver-sb\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.886795 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-ovsdbserver-sb\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.893342 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-dns-svc\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.893992 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-config\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.894134 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-openstack-edpm-ipam\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.894484 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4420782-e4e9-4241-9083-de9ace25d296-ovsdbserver-nb\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:27 crc kubenswrapper[4806]: I1127 10:42:27.945579 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b96j2\" (UniqueName: \"kubernetes.io/projected/c4420782-e4e9-4241-9083-de9ace25d296-kube-api-access-b96j2\") pod \"dnsmasq-dns-667ff9c869-wcdvm\" (UID: \"c4420782-e4e9-4241-9083-de9ace25d296\") " pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.044024 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.125463 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.191903 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npm4s\" (UniqueName: \"kubernetes.io/projected/356b8268-4acb-4ccd-92c5-b085c9fdd884-kube-api-access-npm4s\") pod \"356b8268-4acb-4ccd-92c5-b085c9fdd884\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.191964 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-dns-svc\") pod \"356b8268-4acb-4ccd-92c5-b085c9fdd884\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.192055 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-config\") pod \"356b8268-4acb-4ccd-92c5-b085c9fdd884\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.192163 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-ovsdbserver-nb\") pod \"356b8268-4acb-4ccd-92c5-b085c9fdd884\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.192241 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-ovsdbserver-sb\") pod \"356b8268-4acb-4ccd-92c5-b085c9fdd884\" (UID: \"356b8268-4acb-4ccd-92c5-b085c9fdd884\") " Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.198877 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/356b8268-4acb-4ccd-92c5-b085c9fdd884-kube-api-access-npm4s" (OuterVolumeSpecName: "kube-api-access-npm4s") pod "356b8268-4acb-4ccd-92c5-b085c9fdd884" (UID: "356b8268-4acb-4ccd-92c5-b085c9fdd884"). InnerVolumeSpecName "kube-api-access-npm4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.271221 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "356b8268-4acb-4ccd-92c5-b085c9fdd884" (UID: "356b8268-4acb-4ccd-92c5-b085c9fdd884"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.298372 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npm4s\" (UniqueName: \"kubernetes.io/projected/356b8268-4acb-4ccd-92c5-b085c9fdd884-kube-api-access-npm4s\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.298411 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.309342 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-config" (OuterVolumeSpecName: "config") pod "356b8268-4acb-4ccd-92c5-b085c9fdd884" (UID: "356b8268-4acb-4ccd-92c5-b085c9fdd884"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.330341 4806 generic.go:334] "Generic (PLEG): container finished" podID="356b8268-4acb-4ccd-92c5-b085c9fdd884" containerID="14fde9f23ade1454eddf0c593d445bad7df3fa72e3f58ec0d0de0fc4368be60c" exitCode=0 Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.330412 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" event={"ID":"356b8268-4acb-4ccd-92c5-b085c9fdd884","Type":"ContainerDied","Data":"14fde9f23ade1454eddf0c593d445bad7df3fa72e3f58ec0d0de0fc4368be60c"} Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.330450 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" event={"ID":"356b8268-4acb-4ccd-92c5-b085c9fdd884","Type":"ContainerDied","Data":"78a6503c474989813ccb7e4362e29231fb130d7c88c2187e25ecbf348c2c6368"} Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.330488 4806 scope.go:117] "RemoveContainer" containerID="14fde9f23ade1454eddf0c593d445bad7df3fa72e3f58ec0d0de0fc4368be60c" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.330738 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-dkkxb" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.336128 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "356b8268-4acb-4ccd-92c5-b085c9fdd884" (UID: "356b8268-4acb-4ccd-92c5-b085c9fdd884"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.346097 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "356b8268-4acb-4ccd-92c5-b085c9fdd884" (UID: "356b8268-4acb-4ccd-92c5-b085c9fdd884"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.361906 4806 scope.go:117] "RemoveContainer" containerID="4a6dd3d7cdc811fe1d7ab3a7da7da2163bec8f67f20aa03b9775b8e638e66db7" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.400178 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.400221 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.400417 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/356b8268-4acb-4ccd-92c5-b085c9fdd884-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.402299 4806 scope.go:117] "RemoveContainer" containerID="14fde9f23ade1454eddf0c593d445bad7df3fa72e3f58ec0d0de0fc4368be60c" Nov 27 10:42:28 crc kubenswrapper[4806]: E1127 10:42:28.402994 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14fde9f23ade1454eddf0c593d445bad7df3fa72e3f58ec0d0de0fc4368be60c\": container with ID starting with 14fde9f23ade1454eddf0c593d445bad7df3fa72e3f58ec0d0de0fc4368be60c not found: ID does not exist" containerID="14fde9f23ade1454eddf0c593d445bad7df3fa72e3f58ec0d0de0fc4368be60c" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.403030 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14fde9f23ade1454eddf0c593d445bad7df3fa72e3f58ec0d0de0fc4368be60c"} err="failed to get container status \"14fde9f23ade1454eddf0c593d445bad7df3fa72e3f58ec0d0de0fc4368be60c\": rpc error: code = NotFound desc = could not find container \"14fde9f23ade1454eddf0c593d445bad7df3fa72e3f58ec0d0de0fc4368be60c\": container with ID starting with 14fde9f23ade1454eddf0c593d445bad7df3fa72e3f58ec0d0de0fc4368be60c not found: ID does not exist" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.403062 4806 scope.go:117] "RemoveContainer" containerID="4a6dd3d7cdc811fe1d7ab3a7da7da2163bec8f67f20aa03b9775b8e638e66db7" Nov 27 10:42:28 crc kubenswrapper[4806]: E1127 10:42:28.403646 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a6dd3d7cdc811fe1d7ab3a7da7da2163bec8f67f20aa03b9775b8e638e66db7\": container with ID starting with 4a6dd3d7cdc811fe1d7ab3a7da7da2163bec8f67f20aa03b9775b8e638e66db7 not found: ID does not exist" containerID="4a6dd3d7cdc811fe1d7ab3a7da7da2163bec8f67f20aa03b9775b8e638e66db7" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.403669 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a6dd3d7cdc811fe1d7ab3a7da7da2163bec8f67f20aa03b9775b8e638e66db7"} err="failed to get container status \"4a6dd3d7cdc811fe1d7ab3a7da7da2163bec8f67f20aa03b9775b8e638e66db7\": rpc error: code = NotFound desc = could not find container \"4a6dd3d7cdc811fe1d7ab3a7da7da2163bec8f67f20aa03b9775b8e638e66db7\": container with ID starting with 4a6dd3d7cdc811fe1d7ab3a7da7da2163bec8f67f20aa03b9775b8e638e66db7 not found: ID does not exist" Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.662957 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-667ff9c869-wcdvm"] Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.677125 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-dkkxb"] Nov 27 10:42:28 crc kubenswrapper[4806]: W1127 10:42:28.679950 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4420782_e4e9_4241_9083_de9ace25d296.slice/crio-da2b92e378f5fdfbfb963591ba77182ec56d1f4f01f6d292f6bc66226c9635f7 WatchSource:0}: Error finding container da2b92e378f5fdfbfb963591ba77182ec56d1f4f01f6d292f6bc66226c9635f7: Status 404 returned error can't find the container with id da2b92e378f5fdfbfb963591ba77182ec56d1f4f01f6d292f6bc66226c9635f7 Nov 27 10:42:28 crc kubenswrapper[4806]: I1127 10:42:28.688907 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-dkkxb"] Nov 27 10:42:29 crc kubenswrapper[4806]: I1127 10:42:29.345785 4806 generic.go:334] "Generic (PLEG): container finished" podID="c4420782-e4e9-4241-9083-de9ace25d296" containerID="4fe7de21b84f17cefa82c55564110006104a184c6bf67ceb6b52e2d20c8ee502" exitCode=0 Nov 27 10:42:29 crc kubenswrapper[4806]: I1127 10:42:29.346145 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" event={"ID":"c4420782-e4e9-4241-9083-de9ace25d296","Type":"ContainerDied","Data":"4fe7de21b84f17cefa82c55564110006104a184c6bf67ceb6b52e2d20c8ee502"} Nov 27 10:42:29 crc kubenswrapper[4806]: I1127 10:42:29.346484 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" event={"ID":"c4420782-e4e9-4241-9083-de9ace25d296","Type":"ContainerStarted","Data":"da2b92e378f5fdfbfb963591ba77182ec56d1f4f01f6d292f6bc66226c9635f7"} Nov 27 10:42:30 crc kubenswrapper[4806]: I1127 10:42:30.148294 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="356b8268-4acb-4ccd-92c5-b085c9fdd884" path="/var/lib/kubelet/pods/356b8268-4acb-4ccd-92c5-b085c9fdd884/volumes" Nov 27 10:42:30 crc kubenswrapper[4806]: I1127 10:42:30.359420 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" event={"ID":"c4420782-e4e9-4241-9083-de9ace25d296","Type":"ContainerStarted","Data":"4fa2259576cd47e6699454ed401e1ca39b976c7417c55f32345df8c1109aa811"} Nov 27 10:42:30 crc kubenswrapper[4806]: I1127 10:42:30.360831 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:30 crc kubenswrapper[4806]: I1127 10:42:30.390322 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" podStartSLOduration=3.390296372 podStartE2EDuration="3.390296372s" podCreationTimestamp="2025-11-27 10:42:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:42:30.379759875 +0000 UTC m=+1254.966350639" watchObservedRunningTime="2025-11-27 10:42:30.390296372 +0000 UTC m=+1254.976887136" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.045502 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-667ff9c869-wcdvm" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.158373 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-t2ptf"] Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.158690 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" podUID="615d096c-c354-4ba0-ad6c-c1b25e50033a" containerName="dnsmasq-dns" containerID="cri-o://d770fcdeac8209d8bce8db8ce69c34bccd8e4205cd4473341b20f4502c7eb3fa" gracePeriod=10 Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.445879 4806 generic.go:334] "Generic (PLEG): container finished" podID="615d096c-c354-4ba0-ad6c-c1b25e50033a" containerID="d770fcdeac8209d8bce8db8ce69c34bccd8e4205cd4473341b20f4502c7eb3fa" exitCode=0 Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.445980 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" event={"ID":"615d096c-c354-4ba0-ad6c-c1b25e50033a","Type":"ContainerDied","Data":"d770fcdeac8209d8bce8db8ce69c34bccd8e4205cd4473341b20f4502c7eb3fa"} Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.710355 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.767171 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-ovsdbserver-nb\") pod \"615d096c-c354-4ba0-ad6c-c1b25e50033a\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.767662 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-config\") pod \"615d096c-c354-4ba0-ad6c-c1b25e50033a\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.767703 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-dns-svc\") pod \"615d096c-c354-4ba0-ad6c-c1b25e50033a\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.767870 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-openstack-edpm-ipam\") pod \"615d096c-c354-4ba0-ad6c-c1b25e50033a\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.768001 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hf4w\" (UniqueName: \"kubernetes.io/projected/615d096c-c354-4ba0-ad6c-c1b25e50033a-kube-api-access-6hf4w\") pod \"615d096c-c354-4ba0-ad6c-c1b25e50033a\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.768036 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-ovsdbserver-sb\") pod \"615d096c-c354-4ba0-ad6c-c1b25e50033a\" (UID: \"615d096c-c354-4ba0-ad6c-c1b25e50033a\") " Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.790589 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/615d096c-c354-4ba0-ad6c-c1b25e50033a-kube-api-access-6hf4w" (OuterVolumeSpecName: "kube-api-access-6hf4w") pod "615d096c-c354-4ba0-ad6c-c1b25e50033a" (UID: "615d096c-c354-4ba0-ad6c-c1b25e50033a"). InnerVolumeSpecName "kube-api-access-6hf4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.850893 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "615d096c-c354-4ba0-ad6c-c1b25e50033a" (UID: "615d096c-c354-4ba0-ad6c-c1b25e50033a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.855343 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "615d096c-c354-4ba0-ad6c-c1b25e50033a" (UID: "615d096c-c354-4ba0-ad6c-c1b25e50033a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.865892 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-config" (OuterVolumeSpecName: "config") pod "615d096c-c354-4ba0-ad6c-c1b25e50033a" (UID: "615d096c-c354-4ba0-ad6c-c1b25e50033a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.870389 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hf4w\" (UniqueName: \"kubernetes.io/projected/615d096c-c354-4ba0-ad6c-c1b25e50033a-kube-api-access-6hf4w\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.870514 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.870658 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.870756 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-config\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.880901 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "615d096c-c354-4ba0-ad6c-c1b25e50033a" (UID: "615d096c-c354-4ba0-ad6c-c1b25e50033a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.901213 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "615d096c-c354-4ba0-ad6c-c1b25e50033a" (UID: "615d096c-c354-4ba0-ad6c-c1b25e50033a"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.973692 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:38 crc kubenswrapper[4806]: I1127 10:42:38.973754 4806 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/615d096c-c354-4ba0-ad6c-c1b25e50033a-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 27 10:42:39 crc kubenswrapper[4806]: I1127 10:42:39.458967 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" event={"ID":"615d096c-c354-4ba0-ad6c-c1b25e50033a","Type":"ContainerDied","Data":"2926dcde78f8fc9e108b8f319ab30c09bd51a954f119d7597f206dcf7d5d802e"} Nov 27 10:42:39 crc kubenswrapper[4806]: I1127 10:42:39.459049 4806 scope.go:117] "RemoveContainer" containerID="d770fcdeac8209d8bce8db8ce69c34bccd8e4205cd4473341b20f4502c7eb3fa" Nov 27 10:42:39 crc kubenswrapper[4806]: I1127 10:42:39.459108 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-t2ptf" Nov 27 10:42:39 crc kubenswrapper[4806]: I1127 10:42:39.500689 4806 scope.go:117] "RemoveContainer" containerID="fa72edb9e27c6e415909e9806dac76c58bbeed37d5de6af3b1e2133f8d4dbb94" Nov 27 10:42:39 crc kubenswrapper[4806]: I1127 10:42:39.504402 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-t2ptf"] Nov 27 10:42:39 crc kubenswrapper[4806]: I1127 10:42:39.532959 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-t2ptf"] Nov 27 10:42:40 crc kubenswrapper[4806]: I1127 10:42:40.128068 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="615d096c-c354-4ba0-ad6c-c1b25e50033a" path="/var/lib/kubelet/pods/615d096c-c354-4ba0-ad6c-c1b25e50033a/volumes" Nov 27 10:42:44 crc kubenswrapper[4806]: I1127 10:42:44.450282 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:42:44 crc kubenswrapper[4806]: I1127 10:42:44.451012 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.276555 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t"] Nov 27 10:42:48 crc kubenswrapper[4806]: E1127 10:42:48.278755 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="615d096c-c354-4ba0-ad6c-c1b25e50033a" containerName="init" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.278786 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="615d096c-c354-4ba0-ad6c-c1b25e50033a" containerName="init" Nov 27 10:42:48 crc kubenswrapper[4806]: E1127 10:42:48.278815 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="615d096c-c354-4ba0-ad6c-c1b25e50033a" containerName="dnsmasq-dns" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.278824 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="615d096c-c354-4ba0-ad6c-c1b25e50033a" containerName="dnsmasq-dns" Nov 27 10:42:48 crc kubenswrapper[4806]: E1127 10:42:48.278856 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="356b8268-4acb-4ccd-92c5-b085c9fdd884" containerName="init" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.278867 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="356b8268-4acb-4ccd-92c5-b085c9fdd884" containerName="init" Nov 27 10:42:48 crc kubenswrapper[4806]: E1127 10:42:48.278878 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="356b8268-4acb-4ccd-92c5-b085c9fdd884" containerName="dnsmasq-dns" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.278884 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="356b8268-4acb-4ccd-92c5-b085c9fdd884" containerName="dnsmasq-dns" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.279055 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="615d096c-c354-4ba0-ad6c-c1b25e50033a" containerName="dnsmasq-dns" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.279080 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="356b8268-4acb-4ccd-92c5-b085c9fdd884" containerName="dnsmasq-dns" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.279888 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.282288 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.282775 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.283040 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.291057 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j9ghp" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.297291 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t"] Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.375114 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9862t\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.375412 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9862t\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.375889 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9862t\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.376224 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm6tq\" (UniqueName: \"kubernetes.io/projected/8969ea2b-0bca-42aa-977a-4897c1ce29fb-kube-api-access-cm6tq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9862t\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.478010 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm6tq\" (UniqueName: \"kubernetes.io/projected/8969ea2b-0bca-42aa-977a-4897c1ce29fb-kube-api-access-cm6tq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9862t\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.478131 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9862t\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.478209 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9862t\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.478303 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9862t\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.486092 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9862t\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.494521 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9862t\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.495098 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm6tq\" (UniqueName: \"kubernetes.io/projected/8969ea2b-0bca-42aa-977a-4897c1ce29fb-kube-api-access-cm6tq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9862t\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.498663 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9862t\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:48 crc kubenswrapper[4806]: I1127 10:42:48.602019 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:42:49 crc kubenswrapper[4806]: I1127 10:42:49.172968 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t"] Nov 27 10:42:49 crc kubenswrapper[4806]: I1127 10:42:49.569935 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" event={"ID":"8969ea2b-0bca-42aa-977a-4897c1ce29fb","Type":"ContainerStarted","Data":"ce456e1c48e7e7c65e479e9c11583ba2c9918ad8ab026970bf0eb5ba175fe8fb"} Nov 27 10:42:50 crc kubenswrapper[4806]: I1127 10:42:50.581464 4806 generic.go:334] "Generic (PLEG): container finished" podID="5052338d-2912-432a-a022-aa169c1d0930" containerID="fcb13c64aaa4a1a540bda08dae7b399d51a62b28720239fe07254f98f97c913d" exitCode=0 Nov 27 10:42:50 crc kubenswrapper[4806]: I1127 10:42:50.581517 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5052338d-2912-432a-a022-aa169c1d0930","Type":"ContainerDied","Data":"fcb13c64aaa4a1a540bda08dae7b399d51a62b28720239fe07254f98f97c913d"} Nov 27 10:42:51 crc kubenswrapper[4806]: I1127 10:42:51.594327 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5052338d-2912-432a-a022-aa169c1d0930","Type":"ContainerStarted","Data":"5f85488770019d5216a42fea8aa6197ea26e6da88f6bedc7daaf6cce0e93ad5e"} Nov 27 10:42:51 crc kubenswrapper[4806]: I1127 10:42:51.594868 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 27 10:42:51 crc kubenswrapper[4806]: I1127 10:42:51.601957 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3cdb0a58-54e1-4d11-8240-721304f548dd","Type":"ContainerDied","Data":"5af461ece93a88fa27e8d6a182e6025d312ee415c761bf0fa53683c5adb6df88"} Nov 27 10:42:51 crc kubenswrapper[4806]: I1127 10:42:51.602331 4806 generic.go:334] "Generic (PLEG): container finished" podID="3cdb0a58-54e1-4d11-8240-721304f548dd" containerID="5af461ece93a88fa27e8d6a182e6025d312ee415c761bf0fa53683c5adb6df88" exitCode=0 Nov 27 10:42:51 crc kubenswrapper[4806]: I1127 10:42:51.626698 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.626676446 podStartE2EDuration="38.626676446s" podCreationTimestamp="2025-11-27 10:42:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:42:51.625113294 +0000 UTC m=+1276.211704058" watchObservedRunningTime="2025-11-27 10:42:51.626676446 +0000 UTC m=+1276.213267210" Nov 27 10:42:52 crc kubenswrapper[4806]: I1127 10:42:52.652596 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3cdb0a58-54e1-4d11-8240-721304f548dd","Type":"ContainerStarted","Data":"c1759655094205d4b40e6aa1a14593d97da7e4199a6290651897dd67774d8df8"} Nov 27 10:42:52 crc kubenswrapper[4806]: I1127 10:42:52.653186 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:42:52 crc kubenswrapper[4806]: I1127 10:42:52.683190 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.683172642 podStartE2EDuration="37.683172642s" podCreationTimestamp="2025-11-27 10:42:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 10:42:52.680243972 +0000 UTC m=+1277.266834746" watchObservedRunningTime="2025-11-27 10:42:52.683172642 +0000 UTC m=+1277.269763406" Nov 27 10:43:00 crc kubenswrapper[4806]: I1127 10:43:00.753135 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" event={"ID":"8969ea2b-0bca-42aa-977a-4897c1ce29fb","Type":"ContainerStarted","Data":"9f08469a8e67d553f7235b5fc421d94486dff3ec9bb8f80da62dbbc8176e568b"} Nov 27 10:43:00 crc kubenswrapper[4806]: I1127 10:43:00.775047 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" podStartSLOduration=2.326158813 podStartE2EDuration="12.775026129s" podCreationTimestamp="2025-11-27 10:42:48 +0000 UTC" firstStartedPulling="2025-11-27 10:42:49.178105461 +0000 UTC m=+1273.764696225" lastFinishedPulling="2025-11-27 10:42:59.626972777 +0000 UTC m=+1284.213563541" observedRunningTime="2025-11-27 10:43:00.77393707 +0000 UTC m=+1285.360527834" watchObservedRunningTime="2025-11-27 10:43:00.775026129 +0000 UTC m=+1285.361616903" Nov 27 10:43:04 crc kubenswrapper[4806]: I1127 10:43:04.370572 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 27 10:43:05 crc kubenswrapper[4806]: I1127 10:43:05.558509 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 27 10:43:14 crc kubenswrapper[4806]: I1127 10:43:14.450619 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:43:14 crc kubenswrapper[4806]: I1127 10:43:14.451156 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:43:14 crc kubenswrapper[4806]: I1127 10:43:14.451260 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:43:14 crc kubenswrapper[4806]: I1127 10:43:14.452250 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e9030989914737d800499533c98818930f7560ca5f3c521acf106104ae9ed373"} pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 10:43:14 crc kubenswrapper[4806]: I1127 10:43:14.452358 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" containerID="cri-o://e9030989914737d800499533c98818930f7560ca5f3c521acf106104ae9ed373" gracePeriod=600 Nov 27 10:43:14 crc kubenswrapper[4806]: I1127 10:43:14.965833 4806 generic.go:334] "Generic (PLEG): container finished" podID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerID="e9030989914737d800499533c98818930f7560ca5f3c521acf106104ae9ed373" exitCode=0 Nov 27 10:43:14 crc kubenswrapper[4806]: I1127 10:43:14.965885 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerDied","Data":"e9030989914737d800499533c98818930f7560ca5f3c521acf106104ae9ed373"} Nov 27 10:43:14 crc kubenswrapper[4806]: I1127 10:43:14.966331 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2"} Nov 27 10:43:14 crc kubenswrapper[4806]: I1127 10:43:14.966362 4806 scope.go:117] "RemoveContainer" containerID="5908b1682df3b6a3dc72ea1b0d3756efb806c890ef007bbc5b48b3a697274cc4" Nov 27 10:43:15 crc kubenswrapper[4806]: I1127 10:43:15.980202 4806 generic.go:334] "Generic (PLEG): container finished" podID="8969ea2b-0bca-42aa-977a-4897c1ce29fb" containerID="9f08469a8e67d553f7235b5fc421d94486dff3ec9bb8f80da62dbbc8176e568b" exitCode=0 Nov 27 10:43:15 crc kubenswrapper[4806]: I1127 10:43:15.980613 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" event={"ID":"8969ea2b-0bca-42aa-977a-4897c1ce29fb","Type":"ContainerDied","Data":"9f08469a8e67d553f7235b5fc421d94486dff3ec9bb8f80da62dbbc8176e568b"} Nov 27 10:43:17 crc kubenswrapper[4806]: I1127 10:43:17.432913 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:43:17 crc kubenswrapper[4806]: I1127 10:43:17.555480 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-repo-setup-combined-ca-bundle\") pod \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " Nov 27 10:43:17 crc kubenswrapper[4806]: I1127 10:43:17.555553 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cm6tq\" (UniqueName: \"kubernetes.io/projected/8969ea2b-0bca-42aa-977a-4897c1ce29fb-kube-api-access-cm6tq\") pod \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " Nov 27 10:43:17 crc kubenswrapper[4806]: I1127 10:43:17.555671 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-inventory\") pod \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " Nov 27 10:43:17 crc kubenswrapper[4806]: I1127 10:43:17.555834 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-ssh-key\") pod \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\" (UID: \"8969ea2b-0bca-42aa-977a-4897c1ce29fb\") " Nov 27 10:43:17 crc kubenswrapper[4806]: I1127 10:43:17.567450 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8969ea2b-0bca-42aa-977a-4897c1ce29fb-kube-api-access-cm6tq" (OuterVolumeSpecName: "kube-api-access-cm6tq") pod "8969ea2b-0bca-42aa-977a-4897c1ce29fb" (UID: "8969ea2b-0bca-42aa-977a-4897c1ce29fb"). InnerVolumeSpecName "kube-api-access-cm6tq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:43:17 crc kubenswrapper[4806]: I1127 10:43:17.568564 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "8969ea2b-0bca-42aa-977a-4897c1ce29fb" (UID: "8969ea2b-0bca-42aa-977a-4897c1ce29fb"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:43:17 crc kubenswrapper[4806]: I1127 10:43:17.583577 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8969ea2b-0bca-42aa-977a-4897c1ce29fb" (UID: "8969ea2b-0bca-42aa-977a-4897c1ce29fb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:43:17 crc kubenswrapper[4806]: I1127 10:43:17.593026 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-inventory" (OuterVolumeSpecName: "inventory") pod "8969ea2b-0bca-42aa-977a-4897c1ce29fb" (UID: "8969ea2b-0bca-42aa-977a-4897c1ce29fb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:43:17 crc kubenswrapper[4806]: I1127 10:43:17.658527 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cm6tq\" (UniqueName: \"kubernetes.io/projected/8969ea2b-0bca-42aa-977a-4897c1ce29fb-kube-api-access-cm6tq\") on node \"crc\" DevicePath \"\"" Nov 27 10:43:17 crc kubenswrapper[4806]: I1127 10:43:17.658566 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 10:43:17 crc kubenswrapper[4806]: I1127 10:43:17.658578 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 10:43:17 crc kubenswrapper[4806]: I1127 10:43:17.658589 4806 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8969ea2b-0bca-42aa-977a-4897c1ce29fb-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.009942 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" event={"ID":"8969ea2b-0bca-42aa-977a-4897c1ce29fb","Type":"ContainerDied","Data":"ce456e1c48e7e7c65e479e9c11583ba2c9918ad8ab026970bf0eb5ba175fe8fb"} Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.010004 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce456e1c48e7e7c65e479e9c11583ba2c9918ad8ab026970bf0eb5ba175fe8fb" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.010007 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9862t" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.102743 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx"] Nov 27 10:43:18 crc kubenswrapper[4806]: E1127 10:43:18.103626 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8969ea2b-0bca-42aa-977a-4897c1ce29fb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.103644 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8969ea2b-0bca-42aa-977a-4897c1ce29fb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.103819 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="8969ea2b-0bca-42aa-977a-4897c1ce29fb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.104632 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.107019 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.107030 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.107312 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j9ghp" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.108671 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.112997 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx"] Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.272341 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-992mx\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.272423 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqvnn\" (UniqueName: \"kubernetes.io/projected/4b39d4cf-99ff-486c-bd70-0a614a73539f-kube-api-access-fqvnn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-992mx\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.272463 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-992mx\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.272502 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-992mx\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.374303 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-992mx\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.374377 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-992mx\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.374507 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-992mx\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.374528 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqvnn\" (UniqueName: \"kubernetes.io/projected/4b39d4cf-99ff-486c-bd70-0a614a73539f-kube-api-access-fqvnn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-992mx\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.381965 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-992mx\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.382767 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-992mx\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.382965 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-992mx\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.393987 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqvnn\" (UniqueName: \"kubernetes.io/projected/4b39d4cf-99ff-486c-bd70-0a614a73539f-kube-api-access-fqvnn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-992mx\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:18 crc kubenswrapper[4806]: I1127 10:43:18.427921 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:43:19 crc kubenswrapper[4806]: I1127 10:43:19.075978 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx"] Nov 27 10:43:19 crc kubenswrapper[4806]: W1127 10:43:19.078346 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b39d4cf_99ff_486c_bd70_0a614a73539f.slice/crio-162dcb918c1b19498671619f6d286d45c452c95e22f107c0c0ed68893e248ad1 WatchSource:0}: Error finding container 162dcb918c1b19498671619f6d286d45c452c95e22f107c0c0ed68893e248ad1: Status 404 returned error can't find the container with id 162dcb918c1b19498671619f6d286d45c452c95e22f107c0c0ed68893e248ad1 Nov 27 10:43:20 crc kubenswrapper[4806]: I1127 10:43:20.032716 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" event={"ID":"4b39d4cf-99ff-486c-bd70-0a614a73539f","Type":"ContainerStarted","Data":"bbded33798402386e8a498fdb602b3a775d4fa59b6cba58988020ff158336e1c"} Nov 27 10:43:20 crc kubenswrapper[4806]: I1127 10:43:20.033542 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" event={"ID":"4b39d4cf-99ff-486c-bd70-0a614a73539f","Type":"ContainerStarted","Data":"162dcb918c1b19498671619f6d286d45c452c95e22f107c0c0ed68893e248ad1"} Nov 27 10:43:20 crc kubenswrapper[4806]: I1127 10:43:20.056385 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" podStartSLOduration=1.512463521 podStartE2EDuration="2.056366845s" podCreationTimestamp="2025-11-27 10:43:18 +0000 UTC" firstStartedPulling="2025-11-27 10:43:19.082108098 +0000 UTC m=+1303.668698862" lastFinishedPulling="2025-11-27 10:43:19.626011422 +0000 UTC m=+1304.212602186" observedRunningTime="2025-11-27 10:43:20.052137141 +0000 UTC m=+1304.638727905" watchObservedRunningTime="2025-11-27 10:43:20.056366845 +0000 UTC m=+1304.642957609" Nov 27 10:43:51 crc kubenswrapper[4806]: I1127 10:43:51.801766 4806 scope.go:117] "RemoveContainer" containerID="a30671900ed67563a7771f57e42fb02e9c97d36bfd380b80aad0bcd6d38ca86e" Nov 27 10:44:51 crc kubenswrapper[4806]: I1127 10:44:51.886437 4806 scope.go:117] "RemoveContainer" containerID="b5f34a455734a87a2d12b56355e08ede8ebdacaade0e860817dc1ea7ce12d66b" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.174824 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w"] Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.176596 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.178574 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.179046 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.208022 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w"] Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.257412 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65aae7d9-30f9-44b7-9bb9-159c181a40a5-config-volume\") pod \"collect-profiles-29404005-hg79w\" (UID: \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.257501 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65aae7d9-30f9-44b7-9bb9-159c181a40a5-secret-volume\") pod \"collect-profiles-29404005-hg79w\" (UID: \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.257543 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txf2f\" (UniqueName: \"kubernetes.io/projected/65aae7d9-30f9-44b7-9bb9-159c181a40a5-kube-api-access-txf2f\") pod \"collect-profiles-29404005-hg79w\" (UID: \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.359536 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65aae7d9-30f9-44b7-9bb9-159c181a40a5-config-volume\") pod \"collect-profiles-29404005-hg79w\" (UID: \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.360053 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65aae7d9-30f9-44b7-9bb9-159c181a40a5-secret-volume\") pod \"collect-profiles-29404005-hg79w\" (UID: \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.360217 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txf2f\" (UniqueName: \"kubernetes.io/projected/65aae7d9-30f9-44b7-9bb9-159c181a40a5-kube-api-access-txf2f\") pod \"collect-profiles-29404005-hg79w\" (UID: \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.360546 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65aae7d9-30f9-44b7-9bb9-159c181a40a5-config-volume\") pod \"collect-profiles-29404005-hg79w\" (UID: \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.369339 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65aae7d9-30f9-44b7-9bb9-159c181a40a5-secret-volume\") pod \"collect-profiles-29404005-hg79w\" (UID: \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.376377 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txf2f\" (UniqueName: \"kubernetes.io/projected/65aae7d9-30f9-44b7-9bb9-159c181a40a5-kube-api-access-txf2f\") pod \"collect-profiles-29404005-hg79w\" (UID: \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.499017 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" Nov 27 10:45:00 crc kubenswrapper[4806]: I1127 10:45:00.992530 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w"] Nov 27 10:45:01 crc kubenswrapper[4806]: I1127 10:45:01.161533 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" event={"ID":"65aae7d9-30f9-44b7-9bb9-159c181a40a5","Type":"ContainerStarted","Data":"24d8df5c31ef2a6552c82d4a4cb41723e1348ec3246afd6832811b34681485e3"} Nov 27 10:45:02 crc kubenswrapper[4806]: I1127 10:45:02.172908 4806 generic.go:334] "Generic (PLEG): container finished" podID="65aae7d9-30f9-44b7-9bb9-159c181a40a5" containerID="487898094a090973dd702cedc8f1e0f5a7253de02010c5d7e3e39aae97be6f28" exitCode=0 Nov 27 10:45:02 crc kubenswrapper[4806]: I1127 10:45:02.172966 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" event={"ID":"65aae7d9-30f9-44b7-9bb9-159c181a40a5","Type":"ContainerDied","Data":"487898094a090973dd702cedc8f1e0f5a7253de02010c5d7e3e39aae97be6f28"} Nov 27 10:45:03 crc kubenswrapper[4806]: I1127 10:45:03.563872 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" Nov 27 10:45:03 crc kubenswrapper[4806]: I1127 10:45:03.629784 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txf2f\" (UniqueName: \"kubernetes.io/projected/65aae7d9-30f9-44b7-9bb9-159c181a40a5-kube-api-access-txf2f\") pod \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\" (UID: \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\") " Nov 27 10:45:03 crc kubenswrapper[4806]: I1127 10:45:03.630382 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65aae7d9-30f9-44b7-9bb9-159c181a40a5-config-volume\") pod \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\" (UID: \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\") " Nov 27 10:45:03 crc kubenswrapper[4806]: I1127 10:45:03.630548 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65aae7d9-30f9-44b7-9bb9-159c181a40a5-secret-volume\") pod \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\" (UID: \"65aae7d9-30f9-44b7-9bb9-159c181a40a5\") " Nov 27 10:45:03 crc kubenswrapper[4806]: I1127 10:45:03.631367 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65aae7d9-30f9-44b7-9bb9-159c181a40a5-config-volume" (OuterVolumeSpecName: "config-volume") pod "65aae7d9-30f9-44b7-9bb9-159c181a40a5" (UID: "65aae7d9-30f9-44b7-9bb9-159c181a40a5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 10:45:03 crc kubenswrapper[4806]: I1127 10:45:03.639492 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65aae7d9-30f9-44b7-9bb9-159c181a40a5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "65aae7d9-30f9-44b7-9bb9-159c181a40a5" (UID: "65aae7d9-30f9-44b7-9bb9-159c181a40a5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:45:03 crc kubenswrapper[4806]: I1127 10:45:03.640528 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65aae7d9-30f9-44b7-9bb9-159c181a40a5-kube-api-access-txf2f" (OuterVolumeSpecName: "kube-api-access-txf2f") pod "65aae7d9-30f9-44b7-9bb9-159c181a40a5" (UID: "65aae7d9-30f9-44b7-9bb9-159c181a40a5"). InnerVolumeSpecName "kube-api-access-txf2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:45:03 crc kubenswrapper[4806]: I1127 10:45:03.732960 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65aae7d9-30f9-44b7-9bb9-159c181a40a5-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 10:45:03 crc kubenswrapper[4806]: I1127 10:45:03.732996 4806 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/65aae7d9-30f9-44b7-9bb9-159c181a40a5-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 10:45:03 crc kubenswrapper[4806]: I1127 10:45:03.733008 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txf2f\" (UniqueName: \"kubernetes.io/projected/65aae7d9-30f9-44b7-9bb9-159c181a40a5-kube-api-access-txf2f\") on node \"crc\" DevicePath \"\"" Nov 27 10:45:04 crc kubenswrapper[4806]: I1127 10:45:04.192534 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" event={"ID":"65aae7d9-30f9-44b7-9bb9-159c181a40a5","Type":"ContainerDied","Data":"24d8df5c31ef2a6552c82d4a4cb41723e1348ec3246afd6832811b34681485e3"} Nov 27 10:45:04 crc kubenswrapper[4806]: I1127 10:45:04.192981 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24d8df5c31ef2a6552c82d4a4cb41723e1348ec3246afd6832811b34681485e3" Nov 27 10:45:04 crc kubenswrapper[4806]: I1127 10:45:04.192789 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404005-hg79w" Nov 27 10:45:14 crc kubenswrapper[4806]: I1127 10:45:14.451276 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:45:14 crc kubenswrapper[4806]: I1127 10:45:14.452158 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:45:44 crc kubenswrapper[4806]: I1127 10:45:44.450606 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:45:44 crc kubenswrapper[4806]: I1127 10:45:44.451489 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:46:14 crc kubenswrapper[4806]: I1127 10:46:14.451046 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:46:14 crc kubenswrapper[4806]: I1127 10:46:14.452156 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:46:14 crc kubenswrapper[4806]: I1127 10:46:14.452219 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:46:14 crc kubenswrapper[4806]: I1127 10:46:14.453445 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2"} pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 10:46:14 crc kubenswrapper[4806]: I1127 10:46:14.453530 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" containerID="cri-o://659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" gracePeriod=600 Nov 27 10:46:14 crc kubenswrapper[4806]: E1127 10:46:14.590573 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:46:14 crc kubenswrapper[4806]: I1127 10:46:14.956595 4806 generic.go:334] "Generic (PLEG): container finished" podID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" exitCode=0 Nov 27 10:46:14 crc kubenswrapper[4806]: I1127 10:46:14.956753 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerDied","Data":"659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2"} Nov 27 10:46:14 crc kubenswrapper[4806]: I1127 10:46:14.957000 4806 scope.go:117] "RemoveContainer" containerID="e9030989914737d800499533c98818930f7560ca5f3c521acf106104ae9ed373" Nov 27 10:46:14 crc kubenswrapper[4806]: I1127 10:46:14.957711 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:46:14 crc kubenswrapper[4806]: E1127 10:46:14.957943 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:46:29 crc kubenswrapper[4806]: I1127 10:46:29.116804 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:46:29 crc kubenswrapper[4806]: E1127 10:46:29.117839 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:46:40 crc kubenswrapper[4806]: I1127 10:46:40.117130 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:46:40 crc kubenswrapper[4806]: E1127 10:46:40.118143 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:46:51 crc kubenswrapper[4806]: I1127 10:46:51.988126 4806 scope.go:117] "RemoveContainer" containerID="e4e16d74d4eeb1be7714939610a882a402bb7354f011a3cd781fbde78945e693" Nov 27 10:46:54 crc kubenswrapper[4806]: I1127 10:46:54.116952 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:46:54 crc kubenswrapper[4806]: E1127 10:46:54.117624 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:47:08 crc kubenswrapper[4806]: I1127 10:47:08.117090 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:47:08 crc kubenswrapper[4806]: E1127 10:47:08.119097 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.067745 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-w7d78"] Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.084803 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-c11b-account-create-update-dkh64"] Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.094386 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-qfb6s"] Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.106251 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-3292-account-create-update-p4vfd"] Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.116546 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:47:22 crc kubenswrapper[4806]: E1127 10:47:22.116936 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.125606 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-lkmxz"] Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.125649 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-c11b-account-create-update-dkh64"] Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.132739 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-qfb6s"] Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.139870 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-w7d78"] Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.147337 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-4a42-account-create-update-rm55v"] Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.155556 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-3292-account-create-update-p4vfd"] Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.163737 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-lkmxz"] Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.172134 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-4a42-account-create-update-rm55v"] Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.691693 4806 generic.go:334] "Generic (PLEG): container finished" podID="4b39d4cf-99ff-486c-bd70-0a614a73539f" containerID="bbded33798402386e8a498fdb602b3a775d4fa59b6cba58988020ff158336e1c" exitCode=0 Nov 27 10:47:22 crc kubenswrapper[4806]: I1127 10:47:22.691773 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" event={"ID":"4b39d4cf-99ff-486c-bd70-0a614a73539f","Type":"ContainerDied","Data":"bbded33798402386e8a498fdb602b3a775d4fa59b6cba58988020ff158336e1c"} Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.111674 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.128034 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1020f922-5226-4961-aa56-04c97874b6a6" path="/var/lib/kubelet/pods/1020f922-5226-4961-aa56-04c97874b6a6/volumes" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.129019 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="685cee37-d40c-4ee8-ad87-c67acb12fd60" path="/var/lib/kubelet/pods/685cee37-d40c-4ee8-ad87-c67acb12fd60/volumes" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.129759 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3572cba-65ba-47eb-a98a-ff0c4b3feec8" path="/var/lib/kubelet/pods/a3572cba-65ba-47eb-a98a-ff0c4b3feec8/volumes" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.130487 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b92e4ceb-87fe-4521-b931-9e792f9e4590" path="/var/lib/kubelet/pods/b92e4ceb-87fe-4521-b931-9e792f9e4590/volumes" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.131841 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc617567-5421-4521-8957-4b11667db29e" path="/var/lib/kubelet/pods/fc617567-5421-4521-8957-4b11667db29e/volumes" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.132545 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff5722e8-f491-4a02-836a-f2953334ec7d" path="/var/lib/kubelet/pods/ff5722e8-f491-4a02-836a-f2953334ec7d/volumes" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.264528 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b58hr"] Nov 27 10:47:24 crc kubenswrapper[4806]: E1127 10:47:24.265403 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b39d4cf-99ff-486c-bd70-0a614a73539f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.265514 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b39d4cf-99ff-486c-bd70-0a614a73539f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 27 10:47:24 crc kubenswrapper[4806]: E1127 10:47:24.265625 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65aae7d9-30f9-44b7-9bb9-159c181a40a5" containerName="collect-profiles" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.265700 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="65aae7d9-30f9-44b7-9bb9-159c181a40a5" containerName="collect-profiles" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.266033 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="65aae7d9-30f9-44b7-9bb9-159c181a40a5" containerName="collect-profiles" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.266158 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b39d4cf-99ff-486c-bd70-0a614a73539f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.267990 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.272708 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-ssh-key\") pod \"4b39d4cf-99ff-486c-bd70-0a614a73539f\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.274358 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-bootstrap-combined-ca-bundle\") pod \"4b39d4cf-99ff-486c-bd70-0a614a73539f\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.274816 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-inventory\") pod \"4b39d4cf-99ff-486c-bd70-0a614a73539f\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.300335 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqvnn\" (UniqueName: \"kubernetes.io/projected/4b39d4cf-99ff-486c-bd70-0a614a73539f-kube-api-access-fqvnn\") pod \"4b39d4cf-99ff-486c-bd70-0a614a73539f\" (UID: \"4b39d4cf-99ff-486c-bd70-0a614a73539f\") " Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.301532 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "4b39d4cf-99ff-486c-bd70-0a614a73539f" (UID: "4b39d4cf-99ff-486c-bd70-0a614a73539f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.331876 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b39d4cf-99ff-486c-bd70-0a614a73539f-kube-api-access-fqvnn" (OuterVolumeSpecName: "kube-api-access-fqvnn") pod "4b39d4cf-99ff-486c-bd70-0a614a73539f" (UID: "4b39d4cf-99ff-486c-bd70-0a614a73539f"). InnerVolumeSpecName "kube-api-access-fqvnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.359155 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-catalog-content\") pod \"community-operators-b58hr\" (UID: \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\") " pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.359225 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cflkj\" (UniqueName: \"kubernetes.io/projected/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-kube-api-access-cflkj\") pod \"community-operators-b58hr\" (UID: \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\") " pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.359514 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-utilities\") pod \"community-operators-b58hr\" (UID: \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\") " pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.360527 4806 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.360552 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqvnn\" (UniqueName: \"kubernetes.io/projected/4b39d4cf-99ff-486c-bd70-0a614a73539f-kube-api-access-fqvnn\") on node \"crc\" DevicePath \"\"" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.374697 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4b39d4cf-99ff-486c-bd70-0a614a73539f" (UID: "4b39d4cf-99ff-486c-bd70-0a614a73539f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.383774 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-inventory" (OuterVolumeSpecName: "inventory") pod "4b39d4cf-99ff-486c-bd70-0a614a73539f" (UID: "4b39d4cf-99ff-486c-bd70-0a614a73539f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.394538 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b58hr"] Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.463329 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-catalog-content\") pod \"community-operators-b58hr\" (UID: \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\") " pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.464313 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cflkj\" (UniqueName: \"kubernetes.io/projected/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-kube-api-access-cflkj\") pod \"community-operators-b58hr\" (UID: \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\") " pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.464265 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-catalog-content\") pod \"community-operators-b58hr\" (UID: \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\") " pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.464594 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-utilities\") pod \"community-operators-b58hr\" (UID: \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\") " pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.464912 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-utilities\") pod \"community-operators-b58hr\" (UID: \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\") " pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.465552 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.465594 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b39d4cf-99ff-486c-bd70-0a614a73539f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.485156 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cflkj\" (UniqueName: \"kubernetes.io/projected/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-kube-api-access-cflkj\") pod \"community-operators-b58hr\" (UID: \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\") " pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.713769 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" event={"ID":"4b39d4cf-99ff-486c-bd70-0a614a73539f","Type":"ContainerDied","Data":"162dcb918c1b19498671619f6d286d45c452c95e22f107c0c0ed68893e248ad1"} Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.714111 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="162dcb918c1b19498671619f6d286d45c452c95e22f107c0c0ed68893e248ad1" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.713848 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-992mx" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.745926 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.845262 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp"] Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.846796 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.854903 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.855123 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.855937 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.860358 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j9ghp" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.862427 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp"] Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.983735 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7339d278-f5e4-4731-a70e-bfd3735144b4-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-t58pp\" (UID: \"7339d278-f5e4-4731-a70e-bfd3735144b4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.983889 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hs6s\" (UniqueName: \"kubernetes.io/projected/7339d278-f5e4-4731-a70e-bfd3735144b4-kube-api-access-5hs6s\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-t58pp\" (UID: \"7339d278-f5e4-4731-a70e-bfd3735144b4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" Nov 27 10:47:24 crc kubenswrapper[4806]: I1127 10:47:24.983986 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7339d278-f5e4-4731-a70e-bfd3735144b4-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-t58pp\" (UID: \"7339d278-f5e4-4731-a70e-bfd3735144b4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" Nov 27 10:47:25 crc kubenswrapper[4806]: I1127 10:47:25.085841 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7339d278-f5e4-4731-a70e-bfd3735144b4-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-t58pp\" (UID: \"7339d278-f5e4-4731-a70e-bfd3735144b4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" Nov 27 10:47:25 crc kubenswrapper[4806]: I1127 10:47:25.085910 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7339d278-f5e4-4731-a70e-bfd3735144b4-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-t58pp\" (UID: \"7339d278-f5e4-4731-a70e-bfd3735144b4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" Nov 27 10:47:25 crc kubenswrapper[4806]: I1127 10:47:25.086181 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hs6s\" (UniqueName: \"kubernetes.io/projected/7339d278-f5e4-4731-a70e-bfd3735144b4-kube-api-access-5hs6s\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-t58pp\" (UID: \"7339d278-f5e4-4731-a70e-bfd3735144b4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" Nov 27 10:47:25 crc kubenswrapper[4806]: I1127 10:47:25.095671 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7339d278-f5e4-4731-a70e-bfd3735144b4-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-t58pp\" (UID: \"7339d278-f5e4-4731-a70e-bfd3735144b4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" Nov 27 10:47:25 crc kubenswrapper[4806]: I1127 10:47:25.095770 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7339d278-f5e4-4731-a70e-bfd3735144b4-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-t58pp\" (UID: \"7339d278-f5e4-4731-a70e-bfd3735144b4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" Nov 27 10:47:25 crc kubenswrapper[4806]: I1127 10:47:25.115545 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hs6s\" (UniqueName: \"kubernetes.io/projected/7339d278-f5e4-4731-a70e-bfd3735144b4-kube-api-access-5hs6s\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-t58pp\" (UID: \"7339d278-f5e4-4731-a70e-bfd3735144b4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" Nov 27 10:47:25 crc kubenswrapper[4806]: I1127 10:47:25.169410 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" Nov 27 10:47:25 crc kubenswrapper[4806]: I1127 10:47:25.430024 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b58hr"] Nov 27 10:47:25 crc kubenswrapper[4806]: I1127 10:47:25.725512 4806 generic.go:334] "Generic (PLEG): container finished" podID="6c06d1a7-cfc2-4491-98c4-a148e8c0babe" containerID="ae78cfbed6bb760522028912f6099672e46621eb86e22f01a0d1c67f2efbf177" exitCode=0 Nov 27 10:47:25 crc kubenswrapper[4806]: I1127 10:47:25.725600 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b58hr" event={"ID":"6c06d1a7-cfc2-4491-98c4-a148e8c0babe","Type":"ContainerDied","Data":"ae78cfbed6bb760522028912f6099672e46621eb86e22f01a0d1c67f2efbf177"} Nov 27 10:47:25 crc kubenswrapper[4806]: I1127 10:47:25.725875 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b58hr" event={"ID":"6c06d1a7-cfc2-4491-98c4-a148e8c0babe","Type":"ContainerStarted","Data":"f4ebd46ef1bcfbf77034255ddc4bb9e68a0b1f9c4bb98ea5470d961ff762a5eb"} Nov 27 10:47:25 crc kubenswrapper[4806]: I1127 10:47:25.730545 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 10:47:25 crc kubenswrapper[4806]: W1127 10:47:25.775653 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7339d278_f5e4_4731_a70e_bfd3735144b4.slice/crio-c6c23a15583b92fca3907f32973487f2a65e0593b9f66a31eaff125344cb4655 WatchSource:0}: Error finding container c6c23a15583b92fca3907f32973487f2a65e0593b9f66a31eaff125344cb4655: Status 404 returned error can't find the container with id c6c23a15583b92fca3907f32973487f2a65e0593b9f66a31eaff125344cb4655 Nov 27 10:47:25 crc kubenswrapper[4806]: I1127 10:47:25.778692 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp"] Nov 27 10:47:26 crc kubenswrapper[4806]: I1127 10:47:26.761472 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" event={"ID":"7339d278-f5e4-4731-a70e-bfd3735144b4","Type":"ContainerStarted","Data":"4265d12647fb3839f9f8e3a926c8e407e8e54b4ec36e2f27c1460efd8debebc6"} Nov 27 10:47:26 crc kubenswrapper[4806]: I1127 10:47:26.761880 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" event={"ID":"7339d278-f5e4-4731-a70e-bfd3735144b4","Type":"ContainerStarted","Data":"c6c23a15583b92fca3907f32973487f2a65e0593b9f66a31eaff125344cb4655"} Nov 27 10:47:26 crc kubenswrapper[4806]: I1127 10:47:26.784280 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" podStartSLOduration=2.237063786 podStartE2EDuration="2.784257926s" podCreationTimestamp="2025-11-27 10:47:24 +0000 UTC" firstStartedPulling="2025-11-27 10:47:25.778663324 +0000 UTC m=+1550.365254078" lastFinishedPulling="2025-11-27 10:47:26.325857434 +0000 UTC m=+1550.912448218" observedRunningTime="2025-11-27 10:47:26.779009397 +0000 UTC m=+1551.365600161" watchObservedRunningTime="2025-11-27 10:47:26.784257926 +0000 UTC m=+1551.370848690" Nov 27 10:47:27 crc kubenswrapper[4806]: I1127 10:47:27.770862 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b58hr" event={"ID":"6c06d1a7-cfc2-4491-98c4-a148e8c0babe","Type":"ContainerStarted","Data":"33b6400b56361dc6f0403a6b03d2c54dfa0fad52cad09c92efc8230cd320d110"} Nov 27 10:47:29 crc kubenswrapper[4806]: I1127 10:47:29.794871 4806 generic.go:334] "Generic (PLEG): container finished" podID="6c06d1a7-cfc2-4491-98c4-a148e8c0babe" containerID="33b6400b56361dc6f0403a6b03d2c54dfa0fad52cad09c92efc8230cd320d110" exitCode=0 Nov 27 10:47:29 crc kubenswrapper[4806]: I1127 10:47:29.795302 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b58hr" event={"ID":"6c06d1a7-cfc2-4491-98c4-a148e8c0babe","Type":"ContainerDied","Data":"33b6400b56361dc6f0403a6b03d2c54dfa0fad52cad09c92efc8230cd320d110"} Nov 27 10:47:30 crc kubenswrapper[4806]: I1127 10:47:30.808675 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b58hr" event={"ID":"6c06d1a7-cfc2-4491-98c4-a148e8c0babe","Type":"ContainerStarted","Data":"7cea6ae90c4c8ca6859d4be9b35ebf20495aa9f0be90de2c3e3cf0a1329b9ee8"} Nov 27 10:47:30 crc kubenswrapper[4806]: I1127 10:47:30.837980 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b58hr" podStartSLOduration=2.024154668 podStartE2EDuration="6.837958301s" podCreationTimestamp="2025-11-27 10:47:24 +0000 UTC" firstStartedPulling="2025-11-27 10:47:25.730269793 +0000 UTC m=+1550.316860557" lastFinishedPulling="2025-11-27 10:47:30.544073426 +0000 UTC m=+1555.130664190" observedRunningTime="2025-11-27 10:47:30.831751486 +0000 UTC m=+1555.418342270" watchObservedRunningTime="2025-11-27 10:47:30.837958301 +0000 UTC m=+1555.424549055" Nov 27 10:47:34 crc kubenswrapper[4806]: I1127 10:47:34.117729 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:47:34 crc kubenswrapper[4806]: E1127 10:47:34.118439 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:47:34 crc kubenswrapper[4806]: I1127 10:47:34.746639 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:34 crc kubenswrapper[4806]: I1127 10:47:34.747051 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:34 crc kubenswrapper[4806]: I1127 10:47:34.822601 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:35 crc kubenswrapper[4806]: I1127 10:47:35.939646 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:35 crc kubenswrapper[4806]: I1127 10:47:35.987605 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b58hr"] Nov 27 10:47:37 crc kubenswrapper[4806]: I1127 10:47:37.905507 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b58hr" podUID="6c06d1a7-cfc2-4491-98c4-a148e8c0babe" containerName="registry-server" containerID="cri-o://7cea6ae90c4c8ca6859d4be9b35ebf20495aa9f0be90de2c3e3cf0a1329b9ee8" gracePeriod=2 Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.400313 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.526725 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-catalog-content\") pod \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\" (UID: \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\") " Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.527852 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cflkj\" (UniqueName: \"kubernetes.io/projected/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-kube-api-access-cflkj\") pod \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\" (UID: \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\") " Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.528074 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-utilities\") pod \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\" (UID: \"6c06d1a7-cfc2-4491-98c4-a148e8c0babe\") " Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.528811 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-utilities" (OuterVolumeSpecName: "utilities") pod "6c06d1a7-cfc2-4491-98c4-a148e8c0babe" (UID: "6c06d1a7-cfc2-4491-98c4-a148e8c0babe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.529134 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.535684 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-kube-api-access-cflkj" (OuterVolumeSpecName: "kube-api-access-cflkj") pod "6c06d1a7-cfc2-4491-98c4-a148e8c0babe" (UID: "6c06d1a7-cfc2-4491-98c4-a148e8c0babe"). InnerVolumeSpecName "kube-api-access-cflkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.585588 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c06d1a7-cfc2-4491-98c4-a148e8c0babe" (UID: "6c06d1a7-cfc2-4491-98c4-a148e8c0babe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.632030 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.632078 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cflkj\" (UniqueName: \"kubernetes.io/projected/6c06d1a7-cfc2-4491-98c4-a148e8c0babe-kube-api-access-cflkj\") on node \"crc\" DevicePath \"\"" Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.920371 4806 generic.go:334] "Generic (PLEG): container finished" podID="6c06d1a7-cfc2-4491-98c4-a148e8c0babe" containerID="7cea6ae90c4c8ca6859d4be9b35ebf20495aa9f0be90de2c3e3cf0a1329b9ee8" exitCode=0 Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.920430 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b58hr" event={"ID":"6c06d1a7-cfc2-4491-98c4-a148e8c0babe","Type":"ContainerDied","Data":"7cea6ae90c4c8ca6859d4be9b35ebf20495aa9f0be90de2c3e3cf0a1329b9ee8"} Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.920465 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b58hr" event={"ID":"6c06d1a7-cfc2-4491-98c4-a148e8c0babe","Type":"ContainerDied","Data":"f4ebd46ef1bcfbf77034255ddc4bb9e68a0b1f9c4bb98ea5470d961ff762a5eb"} Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.920486 4806 scope.go:117] "RemoveContainer" containerID="7cea6ae90c4c8ca6859d4be9b35ebf20495aa9f0be90de2c3e3cf0a1329b9ee8" Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.920628 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b58hr" Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.951064 4806 scope.go:117] "RemoveContainer" containerID="33b6400b56361dc6f0403a6b03d2c54dfa0fad52cad09c92efc8230cd320d110" Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.984346 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b58hr"] Nov 27 10:47:38 crc kubenswrapper[4806]: I1127 10:47:38.994772 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b58hr"] Nov 27 10:47:39 crc kubenswrapper[4806]: I1127 10:47:39.005102 4806 scope.go:117] "RemoveContainer" containerID="ae78cfbed6bb760522028912f6099672e46621eb86e22f01a0d1c67f2efbf177" Nov 27 10:47:39 crc kubenswrapper[4806]: I1127 10:47:39.043870 4806 scope.go:117] "RemoveContainer" containerID="7cea6ae90c4c8ca6859d4be9b35ebf20495aa9f0be90de2c3e3cf0a1329b9ee8" Nov 27 10:47:39 crc kubenswrapper[4806]: E1127 10:47:39.044265 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cea6ae90c4c8ca6859d4be9b35ebf20495aa9f0be90de2c3e3cf0a1329b9ee8\": container with ID starting with 7cea6ae90c4c8ca6859d4be9b35ebf20495aa9f0be90de2c3e3cf0a1329b9ee8 not found: ID does not exist" containerID="7cea6ae90c4c8ca6859d4be9b35ebf20495aa9f0be90de2c3e3cf0a1329b9ee8" Nov 27 10:47:39 crc kubenswrapper[4806]: I1127 10:47:39.044294 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cea6ae90c4c8ca6859d4be9b35ebf20495aa9f0be90de2c3e3cf0a1329b9ee8"} err="failed to get container status \"7cea6ae90c4c8ca6859d4be9b35ebf20495aa9f0be90de2c3e3cf0a1329b9ee8\": rpc error: code = NotFound desc = could not find container \"7cea6ae90c4c8ca6859d4be9b35ebf20495aa9f0be90de2c3e3cf0a1329b9ee8\": container with ID starting with 7cea6ae90c4c8ca6859d4be9b35ebf20495aa9f0be90de2c3e3cf0a1329b9ee8 not found: ID does not exist" Nov 27 10:47:39 crc kubenswrapper[4806]: I1127 10:47:39.044316 4806 scope.go:117] "RemoveContainer" containerID="33b6400b56361dc6f0403a6b03d2c54dfa0fad52cad09c92efc8230cd320d110" Nov 27 10:47:39 crc kubenswrapper[4806]: E1127 10:47:39.044656 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33b6400b56361dc6f0403a6b03d2c54dfa0fad52cad09c92efc8230cd320d110\": container with ID starting with 33b6400b56361dc6f0403a6b03d2c54dfa0fad52cad09c92efc8230cd320d110 not found: ID does not exist" containerID="33b6400b56361dc6f0403a6b03d2c54dfa0fad52cad09c92efc8230cd320d110" Nov 27 10:47:39 crc kubenswrapper[4806]: I1127 10:47:39.044722 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33b6400b56361dc6f0403a6b03d2c54dfa0fad52cad09c92efc8230cd320d110"} err="failed to get container status \"33b6400b56361dc6f0403a6b03d2c54dfa0fad52cad09c92efc8230cd320d110\": rpc error: code = NotFound desc = could not find container \"33b6400b56361dc6f0403a6b03d2c54dfa0fad52cad09c92efc8230cd320d110\": container with ID starting with 33b6400b56361dc6f0403a6b03d2c54dfa0fad52cad09c92efc8230cd320d110 not found: ID does not exist" Nov 27 10:47:39 crc kubenswrapper[4806]: I1127 10:47:39.044742 4806 scope.go:117] "RemoveContainer" containerID="ae78cfbed6bb760522028912f6099672e46621eb86e22f01a0d1c67f2efbf177" Nov 27 10:47:39 crc kubenswrapper[4806]: E1127 10:47:39.045015 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae78cfbed6bb760522028912f6099672e46621eb86e22f01a0d1c67f2efbf177\": container with ID starting with ae78cfbed6bb760522028912f6099672e46621eb86e22f01a0d1c67f2efbf177 not found: ID does not exist" containerID="ae78cfbed6bb760522028912f6099672e46621eb86e22f01a0d1c67f2efbf177" Nov 27 10:47:39 crc kubenswrapper[4806]: I1127 10:47:39.045037 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae78cfbed6bb760522028912f6099672e46621eb86e22f01a0d1c67f2efbf177"} err="failed to get container status \"ae78cfbed6bb760522028912f6099672e46621eb86e22f01a0d1c67f2efbf177\": rpc error: code = NotFound desc = could not find container \"ae78cfbed6bb760522028912f6099672e46621eb86e22f01a0d1c67f2efbf177\": container with ID starting with ae78cfbed6bb760522028912f6099672e46621eb86e22f01a0d1c67f2efbf177 not found: ID does not exist" Nov 27 10:47:40 crc kubenswrapper[4806]: I1127 10:47:40.130818 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c06d1a7-cfc2-4491-98c4-a148e8c0babe" path="/var/lib/kubelet/pods/6c06d1a7-cfc2-4491-98c4-a148e8c0babe/volumes" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.202145 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2d5nd"] Nov 27 10:47:44 crc kubenswrapper[4806]: E1127 10:47:44.203070 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c06d1a7-cfc2-4491-98c4-a148e8c0babe" containerName="extract-content" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.203084 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c06d1a7-cfc2-4491-98c4-a148e8c0babe" containerName="extract-content" Nov 27 10:47:44 crc kubenswrapper[4806]: E1127 10:47:44.203110 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c06d1a7-cfc2-4491-98c4-a148e8c0babe" containerName="extract-utilities" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.203118 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c06d1a7-cfc2-4491-98c4-a148e8c0babe" containerName="extract-utilities" Nov 27 10:47:44 crc kubenswrapper[4806]: E1127 10:47:44.203138 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c06d1a7-cfc2-4491-98c4-a148e8c0babe" containerName="registry-server" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.203144 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c06d1a7-cfc2-4491-98c4-a148e8c0babe" containerName="registry-server" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.203352 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c06d1a7-cfc2-4491-98c4-a148e8c0babe" containerName="registry-server" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.204513 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.223920 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2d5nd"] Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.259572 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/525641e9-e2fb-4862-8fe5-ada748dd490e-catalog-content\") pod \"certified-operators-2d5nd\" (UID: \"525641e9-e2fb-4862-8fe5-ada748dd490e\") " pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.259691 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvj8r\" (UniqueName: \"kubernetes.io/projected/525641e9-e2fb-4862-8fe5-ada748dd490e-kube-api-access-vvj8r\") pod \"certified-operators-2d5nd\" (UID: \"525641e9-e2fb-4862-8fe5-ada748dd490e\") " pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.259791 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/525641e9-e2fb-4862-8fe5-ada748dd490e-utilities\") pod \"certified-operators-2d5nd\" (UID: \"525641e9-e2fb-4862-8fe5-ada748dd490e\") " pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.362460 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/525641e9-e2fb-4862-8fe5-ada748dd490e-utilities\") pod \"certified-operators-2d5nd\" (UID: \"525641e9-e2fb-4862-8fe5-ada748dd490e\") " pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.362587 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/525641e9-e2fb-4862-8fe5-ada748dd490e-catalog-content\") pod \"certified-operators-2d5nd\" (UID: \"525641e9-e2fb-4862-8fe5-ada748dd490e\") " pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.362655 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvj8r\" (UniqueName: \"kubernetes.io/projected/525641e9-e2fb-4862-8fe5-ada748dd490e-kube-api-access-vvj8r\") pod \"certified-operators-2d5nd\" (UID: \"525641e9-e2fb-4862-8fe5-ada748dd490e\") " pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.363152 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/525641e9-e2fb-4862-8fe5-ada748dd490e-utilities\") pod \"certified-operators-2d5nd\" (UID: \"525641e9-e2fb-4862-8fe5-ada748dd490e\") " pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.363444 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/525641e9-e2fb-4862-8fe5-ada748dd490e-catalog-content\") pod \"certified-operators-2d5nd\" (UID: \"525641e9-e2fb-4862-8fe5-ada748dd490e\") " pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.394728 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvj8r\" (UniqueName: \"kubernetes.io/projected/525641e9-e2fb-4862-8fe5-ada748dd490e-kube-api-access-vvj8r\") pod \"certified-operators-2d5nd\" (UID: \"525641e9-e2fb-4862-8fe5-ada748dd490e\") " pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:44 crc kubenswrapper[4806]: I1127 10:47:44.534780 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:45 crc kubenswrapper[4806]: I1127 10:47:45.073364 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2d5nd"] Nov 27 10:47:46 crc kubenswrapper[4806]: I1127 10:47:46.006021 4806 generic.go:334] "Generic (PLEG): container finished" podID="525641e9-e2fb-4862-8fe5-ada748dd490e" containerID="d979435066cfd1307ca9d95c4e85121569802c877e3fb95859ffb1744d894e40" exitCode=0 Nov 27 10:47:46 crc kubenswrapper[4806]: I1127 10:47:46.006128 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2d5nd" event={"ID":"525641e9-e2fb-4862-8fe5-ada748dd490e","Type":"ContainerDied","Data":"d979435066cfd1307ca9d95c4e85121569802c877e3fb95859ffb1744d894e40"} Nov 27 10:47:46 crc kubenswrapper[4806]: I1127 10:47:46.007042 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2d5nd" event={"ID":"525641e9-e2fb-4862-8fe5-ada748dd490e","Type":"ContainerStarted","Data":"1c2d8300c2b57205fbd5cc911f2f0e0d9d6c190f10324f4e0ad3f77649ae8220"} Nov 27 10:47:48 crc kubenswrapper[4806]: I1127 10:47:48.117460 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:47:48 crc kubenswrapper[4806]: E1127 10:47:48.118394 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:47:52 crc kubenswrapper[4806]: I1127 10:47:52.048573 4806 scope.go:117] "RemoveContainer" containerID="83a572048feb1595316cc2d16cd603da84a3d7b676e9d80a2f4f7c58922bbcf1" Nov 27 10:47:52 crc kubenswrapper[4806]: I1127 10:47:52.075515 4806 scope.go:117] "RemoveContainer" containerID="cf7772a86778265847fd4a67a69bd0f1b2072cd2f2fee80e3b9d73606dd7a0d6" Nov 27 10:47:52 crc kubenswrapper[4806]: I1127 10:47:52.092425 4806 generic.go:334] "Generic (PLEG): container finished" podID="525641e9-e2fb-4862-8fe5-ada748dd490e" containerID="b351fa916f6e22a99d10f49429207b2d8627758479f434347ff0894e7165fd80" exitCode=0 Nov 27 10:47:52 crc kubenswrapper[4806]: I1127 10:47:52.092474 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2d5nd" event={"ID":"525641e9-e2fb-4862-8fe5-ada748dd490e","Type":"ContainerDied","Data":"b351fa916f6e22a99d10f49429207b2d8627758479f434347ff0894e7165fd80"} Nov 27 10:47:52 crc kubenswrapper[4806]: I1127 10:47:52.145775 4806 scope.go:117] "RemoveContainer" containerID="2abd910320dc8930648b10301139f8fa31ceee0dca10e90166d2952768870271" Nov 27 10:47:52 crc kubenswrapper[4806]: I1127 10:47:52.191612 4806 scope.go:117] "RemoveContainer" containerID="f3ab201c5083e96c309a8526d027d7add5ab4d4266af330b19b20b947a743221" Nov 27 10:47:52 crc kubenswrapper[4806]: I1127 10:47:52.237667 4806 scope.go:117] "RemoveContainer" containerID="b3bb5f91d43d19f6b89a5a7e99071c2ba7ecca58bdf586c7c400bb174c06088e" Nov 27 10:47:52 crc kubenswrapper[4806]: I1127 10:47:52.277386 4806 scope.go:117] "RemoveContainer" containerID="d6b07eb372a44b87726e1e2e0264485b33b0c4773996c678d1f60e99f2f60d6a" Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.070552 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-3071-account-create-update-2g8jt"] Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.085147 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-rhbrc"] Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.110577 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-3a87-account-create-update-8zm2q"] Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.120928 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2d5nd" event={"ID":"525641e9-e2fb-4862-8fe5-ada748dd490e","Type":"ContainerStarted","Data":"5fe5a046e0a152773d6392d4f750b4c8c173f68135acaab1eb0e266d440021b1"} Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.121829 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-f488s"] Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.130459 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-xptpp"] Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.144537 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-1c99-account-create-update-6mm4j"] Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.151748 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-xptpp"] Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.165370 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-f488s"] Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.174371 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-3071-account-create-update-2g8jt"] Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.182449 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-1c99-account-create-update-6mm4j"] Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.191303 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-3a87-account-create-update-8zm2q"] Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.200190 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-rhbrc"] Nov 27 10:47:53 crc kubenswrapper[4806]: I1127 10:47:53.209698 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2d5nd" podStartSLOduration=2.6170428340000003 podStartE2EDuration="9.209671816s" podCreationTimestamp="2025-11-27 10:47:44 +0000 UTC" firstStartedPulling="2025-11-27 10:47:46.010611395 +0000 UTC m=+1570.597202159" lastFinishedPulling="2025-11-27 10:47:52.603240377 +0000 UTC m=+1577.189831141" observedRunningTime="2025-11-27 10:47:53.144615591 +0000 UTC m=+1577.731206345" watchObservedRunningTime="2025-11-27 10:47:53.209671816 +0000 UTC m=+1577.796262590" Nov 27 10:47:54 crc kubenswrapper[4806]: I1127 10:47:54.128755 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="679580af-7a44-485f-8b79-f79b702d85eb" path="/var/lib/kubelet/pods/679580af-7a44-485f-8b79-f79b702d85eb/volumes" Nov 27 10:47:54 crc kubenswrapper[4806]: I1127 10:47:54.129709 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82baaafd-2458-40d0-b859-8171f9a6db0d" path="/var/lib/kubelet/pods/82baaafd-2458-40d0-b859-8171f9a6db0d/volumes" Nov 27 10:47:54 crc kubenswrapper[4806]: I1127 10:47:54.130206 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3e71c8a-a010-49f6-9e79-c868ec3243d9" path="/var/lib/kubelet/pods/a3e71c8a-a010-49f6-9e79-c868ec3243d9/volumes" Nov 27 10:47:54 crc kubenswrapper[4806]: I1127 10:47:54.130724 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9270118-f53f-4b38-98e6-03469db9302d" path="/var/lib/kubelet/pods/a9270118-f53f-4b38-98e6-03469db9302d/volumes" Nov 27 10:47:54 crc kubenswrapper[4806]: I1127 10:47:54.131835 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df636244-5318-44a8-857b-2c68af37b10a" path="/var/lib/kubelet/pods/df636244-5318-44a8-857b-2c68af37b10a/volumes" Nov 27 10:47:54 crc kubenswrapper[4806]: I1127 10:47:54.132422 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7143fe2-d8ae-4d55-a11c-edb06cf922f7" path="/var/lib/kubelet/pods/f7143fe2-d8ae-4d55-a11c-edb06cf922f7/volumes" Nov 27 10:47:54 crc kubenswrapper[4806]: I1127 10:47:54.536785 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:54 crc kubenswrapper[4806]: I1127 10:47:54.536845 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:54 crc kubenswrapper[4806]: I1127 10:47:54.587102 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:47:58 crc kubenswrapper[4806]: I1127 10:47:58.046682 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-6vcz5"] Nov 27 10:47:58 crc kubenswrapper[4806]: I1127 10:47:58.057620 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-6vcz5"] Nov 27 10:47:58 crc kubenswrapper[4806]: I1127 10:47:58.128812 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7484e21-2c80-49d6-a9ec-6489bc40ec4f" path="/var/lib/kubelet/pods/c7484e21-2c80-49d6-a9ec-6489bc40ec4f/volumes" Nov 27 10:47:59 crc kubenswrapper[4806]: I1127 10:47:59.031750 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-h8mc6"] Nov 27 10:47:59 crc kubenswrapper[4806]: I1127 10:47:59.043729 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-h8mc6"] Nov 27 10:48:00 crc kubenswrapper[4806]: I1127 10:48:00.128942 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aefc973-2bd6-449e-a7e7-afd6a7915c7a" path="/var/lib/kubelet/pods/9aefc973-2bd6-449e-a7e7-afd6a7915c7a/volumes" Nov 27 10:48:02 crc kubenswrapper[4806]: I1127 10:48:02.117311 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:48:02 crc kubenswrapper[4806]: E1127 10:48:02.117964 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:48:04 crc kubenswrapper[4806]: I1127 10:48:04.587429 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2d5nd" Nov 27 10:48:04 crc kubenswrapper[4806]: I1127 10:48:04.657999 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2d5nd"] Nov 27 10:48:04 crc kubenswrapper[4806]: I1127 10:48:04.761943 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wfxvq"] Nov 27 10:48:04 crc kubenswrapper[4806]: I1127 10:48:04.762523 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wfxvq" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" containerName="registry-server" containerID="cri-o://5eb7db305ecb246b3fd9bafaac53f10afac8a86ea08e3245141892892620d3f3" gracePeriod=2 Nov 27 10:48:05 crc kubenswrapper[4806]: I1127 10:48:05.253321 4806 generic.go:334] "Generic (PLEG): container finished" podID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" containerID="5eb7db305ecb246b3fd9bafaac53f10afac8a86ea08e3245141892892620d3f3" exitCode=0 Nov 27 10:48:05 crc kubenswrapper[4806]: I1127 10:48:05.253381 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfxvq" event={"ID":"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3","Type":"ContainerDied","Data":"5eb7db305ecb246b3fd9bafaac53f10afac8a86ea08e3245141892892620d3f3"} Nov 27 10:48:05 crc kubenswrapper[4806]: I1127 10:48:05.864799 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.027534 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-catalog-content\") pod \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\" (UID: \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\") " Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.027677 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88rp5\" (UniqueName: \"kubernetes.io/projected/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-kube-api-access-88rp5\") pod \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\" (UID: \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\") " Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.027847 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-utilities\") pod \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\" (UID: \"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3\") " Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.028242 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-utilities" (OuterVolumeSpecName: "utilities") pod "dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" (UID: "dcdfc5eb-7543-43b1-bbb5-90734e6f85a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.028462 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.049506 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-kube-api-access-88rp5" (OuterVolumeSpecName: "kube-api-access-88rp5") pod "dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" (UID: "dcdfc5eb-7543-43b1-bbb5-90734e6f85a3"). InnerVolumeSpecName "kube-api-access-88rp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.090848 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" (UID: "dcdfc5eb-7543-43b1-bbb5-90734e6f85a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.130563 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.130619 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88rp5\" (UniqueName: \"kubernetes.io/projected/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3-kube-api-access-88rp5\") on node \"crc\" DevicePath \"\"" Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.265982 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfxvq" event={"ID":"dcdfc5eb-7543-43b1-bbb5-90734e6f85a3","Type":"ContainerDied","Data":"3076fadda0d099ff25c6341026a33f7f0f377f5020968fb1a91de7fa56201a1b"} Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.266422 4806 scope.go:117] "RemoveContainer" containerID="5eb7db305ecb246b3fd9bafaac53f10afac8a86ea08e3245141892892620d3f3" Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.266167 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfxvq" Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.293769 4806 scope.go:117] "RemoveContainer" containerID="d8481b180e70933acad600bdc4e2e433f91c609578f9d03677cfa267f21936f5" Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.308454 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wfxvq"] Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.321316 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wfxvq"] Nov 27 10:48:06 crc kubenswrapper[4806]: I1127 10:48:06.358001 4806 scope.go:117] "RemoveContainer" containerID="94e79ff145bb937ccf9506389b53e3a77f89371f64e7c94d1ddf76d5f32220e6" Nov 27 10:48:08 crc kubenswrapper[4806]: I1127 10:48:08.137004 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" path="/var/lib/kubelet/pods/dcdfc5eb-7543-43b1-bbb5-90734e6f85a3/volumes" Nov 27 10:48:17 crc kubenswrapper[4806]: I1127 10:48:17.117348 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:48:17 crc kubenswrapper[4806]: E1127 10:48:17.118318 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:48:29 crc kubenswrapper[4806]: I1127 10:48:29.048691 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-h7fbq"] Nov 27 10:48:29 crc kubenswrapper[4806]: I1127 10:48:29.056366 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-h7fbq"] Nov 27 10:48:29 crc kubenswrapper[4806]: I1127 10:48:29.117746 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:48:29 crc kubenswrapper[4806]: E1127 10:48:29.118067 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:48:30 crc kubenswrapper[4806]: I1127 10:48:30.137115 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94086dd9-9626-49a4-ad72-af183dc0a161" path="/var/lib/kubelet/pods/94086dd9-9626-49a4-ad72-af183dc0a161/volumes" Nov 27 10:48:38 crc kubenswrapper[4806]: I1127 10:48:38.042137 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jkq5l"] Nov 27 10:48:38 crc kubenswrapper[4806]: I1127 10:48:38.053947 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jkq5l"] Nov 27 10:48:38 crc kubenswrapper[4806]: I1127 10:48:38.069358 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-hz79k"] Nov 27 10:48:38 crc kubenswrapper[4806]: I1127 10:48:38.083853 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-hz79k"] Nov 27 10:48:38 crc kubenswrapper[4806]: I1127 10:48:38.136601 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b289081e-bb54-4046-9e62-37e47cacca99" path="/var/lib/kubelet/pods/b289081e-bb54-4046-9e62-37e47cacca99/volumes" Nov 27 10:48:38 crc kubenswrapper[4806]: I1127 10:48:38.137291 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c109a218-41c4-4872-9fee-fbba98bb0143" path="/var/lib/kubelet/pods/c109a218-41c4-4872-9fee-fbba98bb0143/volumes" Nov 27 10:48:42 crc kubenswrapper[4806]: I1127 10:48:42.119060 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:48:42 crc kubenswrapper[4806]: E1127 10:48:42.121025 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.030563 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-9vkqx"] Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.037977 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-9vkqx"] Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.127216 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11b83e17-71ca-4f23-ad17-950a5de85a12" path="/var/lib/kubelet/pods/11b83e17-71ca-4f23-ad17-950a5de85a12/volumes" Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.461134 4806 scope.go:117] "RemoveContainer" containerID="550e722b7e0b4143095a89bd42da2fc92a7788d6fcc069823e4b4bd9e326ee21" Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.498350 4806 scope.go:117] "RemoveContainer" containerID="884656ef95050d07810144fe740ae52ba53cbdf5fa5c5756fdba61f8c93ed143" Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.558852 4806 scope.go:117] "RemoveContainer" containerID="d2a5eda0802027071d1e68bf8b7f9de6516a5df948f774197491684a44f05028" Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.623491 4806 scope.go:117] "RemoveContainer" containerID="35b4ed1f1f67e783c6ecb121e37e9be0153a83bf8984f252587b15c9d553781c" Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.673809 4806 scope.go:117] "RemoveContainer" containerID="4f96c0b3875c99592487a94e9960ec45541a81e5c5dcb5a45a909916cd0e015a" Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.702667 4806 scope.go:117] "RemoveContainer" containerID="40b85bffaf47fe48ed4844f796570b481a16e94be64512d19ce2624d90183d17" Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.766580 4806 scope.go:117] "RemoveContainer" containerID="eb8d3083c3fa65165665e80e3941c7c11b2e95bac5b598a421fbb172b62015b2" Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.789124 4806 scope.go:117] "RemoveContainer" containerID="f6b40b2335c910e068c63dab1fac7deb316f1213e1f22109ff58b5a9dabe6b03" Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.828296 4806 scope.go:117] "RemoveContainer" containerID="856d1cbfd076b861d6ec0311109278411b8f03ca997a7f6e4b585020b52cb83f" Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.872704 4806 scope.go:117] "RemoveContainer" containerID="4db8338c12c4c2627d7be068f50323ac9588babf56f5307588154ee1b2d68d02" Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.910909 4806 scope.go:117] "RemoveContainer" containerID="85afbb8a4561fd7353250eb568d0df936c03dc313d30b9e1f4ce22eaaa1ae12e" Nov 27 10:48:52 crc kubenswrapper[4806]: I1127 10:48:52.938546 4806 scope.go:117] "RemoveContainer" containerID="b4a0eb7d0744a08010adc745dcfacc358029c1457f0b148d283dba18fe7bcd90" Nov 27 10:48:53 crc kubenswrapper[4806]: I1127 10:48:53.065874 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-cwqkj"] Nov 27 10:48:53 crc kubenswrapper[4806]: I1127 10:48:53.116337 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-cwqkj"] Nov 27 10:48:54 crc kubenswrapper[4806]: I1127 10:48:54.131847 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="552d400b-d5b2-40c8-9654-bfc70ab5bab6" path="/var/lib/kubelet/pods/552d400b-d5b2-40c8-9654-bfc70ab5bab6/volumes" Nov 27 10:48:55 crc kubenswrapper[4806]: I1127 10:48:55.116998 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:48:55 crc kubenswrapper[4806]: E1127 10:48:55.117375 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:49:02 crc kubenswrapper[4806]: I1127 10:49:02.912024 4806 generic.go:334] "Generic (PLEG): container finished" podID="7339d278-f5e4-4731-a70e-bfd3735144b4" containerID="4265d12647fb3839f9f8e3a926c8e407e8e54b4ec36e2f27c1460efd8debebc6" exitCode=0 Nov 27 10:49:02 crc kubenswrapper[4806]: I1127 10:49:02.912110 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" event={"ID":"7339d278-f5e4-4731-a70e-bfd3735144b4","Type":"ContainerDied","Data":"4265d12647fb3839f9f8e3a926c8e407e8e54b4ec36e2f27c1460efd8debebc6"} Nov 27 10:49:04 crc kubenswrapper[4806]: I1127 10:49:04.362611 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" Nov 27 10:49:04 crc kubenswrapper[4806]: I1127 10:49:04.481967 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7339d278-f5e4-4731-a70e-bfd3735144b4-inventory\") pod \"7339d278-f5e4-4731-a70e-bfd3735144b4\" (UID: \"7339d278-f5e4-4731-a70e-bfd3735144b4\") " Nov 27 10:49:04 crc kubenswrapper[4806]: I1127 10:49:04.482218 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7339d278-f5e4-4731-a70e-bfd3735144b4-ssh-key\") pod \"7339d278-f5e4-4731-a70e-bfd3735144b4\" (UID: \"7339d278-f5e4-4731-a70e-bfd3735144b4\") " Nov 27 10:49:04 crc kubenswrapper[4806]: I1127 10:49:04.482266 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hs6s\" (UniqueName: \"kubernetes.io/projected/7339d278-f5e4-4731-a70e-bfd3735144b4-kube-api-access-5hs6s\") pod \"7339d278-f5e4-4731-a70e-bfd3735144b4\" (UID: \"7339d278-f5e4-4731-a70e-bfd3735144b4\") " Nov 27 10:49:04 crc kubenswrapper[4806]: I1127 10:49:04.490024 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7339d278-f5e4-4731-a70e-bfd3735144b4-kube-api-access-5hs6s" (OuterVolumeSpecName: "kube-api-access-5hs6s") pod "7339d278-f5e4-4731-a70e-bfd3735144b4" (UID: "7339d278-f5e4-4731-a70e-bfd3735144b4"). InnerVolumeSpecName "kube-api-access-5hs6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:49:04 crc kubenswrapper[4806]: I1127 10:49:04.518952 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7339d278-f5e4-4731-a70e-bfd3735144b4-inventory" (OuterVolumeSpecName: "inventory") pod "7339d278-f5e4-4731-a70e-bfd3735144b4" (UID: "7339d278-f5e4-4731-a70e-bfd3735144b4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:49:04 crc kubenswrapper[4806]: I1127 10:49:04.557508 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7339d278-f5e4-4731-a70e-bfd3735144b4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7339d278-f5e4-4731-a70e-bfd3735144b4" (UID: "7339d278-f5e4-4731-a70e-bfd3735144b4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:49:04 crc kubenswrapper[4806]: I1127 10:49:04.585584 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7339d278-f5e4-4731-a70e-bfd3735144b4-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 10:49:04 crc kubenswrapper[4806]: I1127 10:49:04.585627 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7339d278-f5e4-4731-a70e-bfd3735144b4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 10:49:04 crc kubenswrapper[4806]: I1127 10:49:04.585641 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hs6s\" (UniqueName: \"kubernetes.io/projected/7339d278-f5e4-4731-a70e-bfd3735144b4-kube-api-access-5hs6s\") on node \"crc\" DevicePath \"\"" Nov 27 10:49:04 crc kubenswrapper[4806]: I1127 10:49:04.931609 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" event={"ID":"7339d278-f5e4-4731-a70e-bfd3735144b4","Type":"ContainerDied","Data":"c6c23a15583b92fca3907f32973487f2a65e0593b9f66a31eaff125344cb4655"} Nov 27 10:49:04 crc kubenswrapper[4806]: I1127 10:49:04.931900 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6c23a15583b92fca3907f32973487f2a65e0593b9f66a31eaff125344cb4655" Nov 27 10:49:04 crc kubenswrapper[4806]: I1127 10:49:04.931688 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-t58pp" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.029067 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb"] Nov 27 10:49:05 crc kubenswrapper[4806]: E1127 10:49:05.029495 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7339d278-f5e4-4731-a70e-bfd3735144b4" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.029515 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7339d278-f5e4-4731-a70e-bfd3735144b4" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 27 10:49:05 crc kubenswrapper[4806]: E1127 10:49:05.029529 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" containerName="registry-server" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.029535 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" containerName="registry-server" Nov 27 10:49:05 crc kubenswrapper[4806]: E1127 10:49:05.029549 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" containerName="extract-utilities" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.029555 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" containerName="extract-utilities" Nov 27 10:49:05 crc kubenswrapper[4806]: E1127 10:49:05.029581 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" containerName="extract-content" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.029587 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" containerName="extract-content" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.029785 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="7339d278-f5e4-4731-a70e-bfd3735144b4" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.029797 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcdfc5eb-7543-43b1-bbb5-90734e6f85a3" containerName="registry-server" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.030527 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.033183 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.033500 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.033921 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.035397 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j9ghp" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.050869 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb"] Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.197679 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7197a402-7e1b-450a-ba53-448bf843e6f6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb\" (UID: \"7197a402-7e1b-450a-ba53-448bf843e6f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.197770 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7197a402-7e1b-450a-ba53-448bf843e6f6-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb\" (UID: \"7197a402-7e1b-450a-ba53-448bf843e6f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.197805 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngj9q\" (UniqueName: \"kubernetes.io/projected/7197a402-7e1b-450a-ba53-448bf843e6f6-kube-api-access-ngj9q\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb\" (UID: \"7197a402-7e1b-450a-ba53-448bf843e6f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.300180 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7197a402-7e1b-450a-ba53-448bf843e6f6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb\" (UID: \"7197a402-7e1b-450a-ba53-448bf843e6f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.300286 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7197a402-7e1b-450a-ba53-448bf843e6f6-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb\" (UID: \"7197a402-7e1b-450a-ba53-448bf843e6f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.300332 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngj9q\" (UniqueName: \"kubernetes.io/projected/7197a402-7e1b-450a-ba53-448bf843e6f6-kube-api-access-ngj9q\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb\" (UID: \"7197a402-7e1b-450a-ba53-448bf843e6f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.306038 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7197a402-7e1b-450a-ba53-448bf843e6f6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb\" (UID: \"7197a402-7e1b-450a-ba53-448bf843e6f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.311077 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7197a402-7e1b-450a-ba53-448bf843e6f6-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb\" (UID: \"7197a402-7e1b-450a-ba53-448bf843e6f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.318839 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngj9q\" (UniqueName: \"kubernetes.io/projected/7197a402-7e1b-450a-ba53-448bf843e6f6-kube-api-access-ngj9q\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb\" (UID: \"7197a402-7e1b-450a-ba53-448bf843e6f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.352658 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" Nov 27 10:49:05 crc kubenswrapper[4806]: I1127 10:49:05.940214 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb"] Nov 27 10:49:06 crc kubenswrapper[4806]: I1127 10:49:06.948835 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" event={"ID":"7197a402-7e1b-450a-ba53-448bf843e6f6","Type":"ContainerStarted","Data":"8e3ff9e8e392f6955ac13a929a66e94d805ed5863b366ab988de72da2f9c4e14"} Nov 27 10:49:06 crc kubenswrapper[4806]: I1127 10:49:06.948884 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" event={"ID":"7197a402-7e1b-450a-ba53-448bf843e6f6","Type":"ContainerStarted","Data":"e1342a8954d97f3a51d897eaf8d586818a5afe974f35419b010444155e0e12bf"} Nov 27 10:49:06 crc kubenswrapper[4806]: I1127 10:49:06.972221 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" podStartSLOduration=1.294081901 podStartE2EDuration="1.972195783s" podCreationTimestamp="2025-11-27 10:49:05 +0000 UTC" firstStartedPulling="2025-11-27 10:49:05.970216557 +0000 UTC m=+1650.556807321" lastFinishedPulling="2025-11-27 10:49:06.648330439 +0000 UTC m=+1651.234921203" observedRunningTime="2025-11-27 10:49:06.962942057 +0000 UTC m=+1651.549532821" watchObservedRunningTime="2025-11-27 10:49:06.972195783 +0000 UTC m=+1651.558786547" Nov 27 10:49:09 crc kubenswrapper[4806]: I1127 10:49:09.116586 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:49:09 crc kubenswrapper[4806]: E1127 10:49:09.117260 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:49:13 crc kubenswrapper[4806]: I1127 10:49:13.004068 4806 generic.go:334] "Generic (PLEG): container finished" podID="7197a402-7e1b-450a-ba53-448bf843e6f6" containerID="8e3ff9e8e392f6955ac13a929a66e94d805ed5863b366ab988de72da2f9c4e14" exitCode=0 Nov 27 10:49:13 crc kubenswrapper[4806]: I1127 10:49:13.004154 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" event={"ID":"7197a402-7e1b-450a-ba53-448bf843e6f6","Type":"ContainerDied","Data":"8e3ff9e8e392f6955ac13a929a66e94d805ed5863b366ab988de72da2f9c4e14"} Nov 27 10:49:14 crc kubenswrapper[4806]: I1127 10:49:14.412190 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" Nov 27 10:49:14 crc kubenswrapper[4806]: I1127 10:49:14.587960 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7197a402-7e1b-450a-ba53-448bf843e6f6-ssh-key\") pod \"7197a402-7e1b-450a-ba53-448bf843e6f6\" (UID: \"7197a402-7e1b-450a-ba53-448bf843e6f6\") " Nov 27 10:49:14 crc kubenswrapper[4806]: I1127 10:49:14.588256 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7197a402-7e1b-450a-ba53-448bf843e6f6-inventory\") pod \"7197a402-7e1b-450a-ba53-448bf843e6f6\" (UID: \"7197a402-7e1b-450a-ba53-448bf843e6f6\") " Nov 27 10:49:14 crc kubenswrapper[4806]: I1127 10:49:14.588292 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngj9q\" (UniqueName: \"kubernetes.io/projected/7197a402-7e1b-450a-ba53-448bf843e6f6-kube-api-access-ngj9q\") pod \"7197a402-7e1b-450a-ba53-448bf843e6f6\" (UID: \"7197a402-7e1b-450a-ba53-448bf843e6f6\") " Nov 27 10:49:14 crc kubenswrapper[4806]: I1127 10:49:14.606693 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7197a402-7e1b-450a-ba53-448bf843e6f6-kube-api-access-ngj9q" (OuterVolumeSpecName: "kube-api-access-ngj9q") pod "7197a402-7e1b-450a-ba53-448bf843e6f6" (UID: "7197a402-7e1b-450a-ba53-448bf843e6f6"). InnerVolumeSpecName "kube-api-access-ngj9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:49:14 crc kubenswrapper[4806]: I1127 10:49:14.614394 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7197a402-7e1b-450a-ba53-448bf843e6f6-inventory" (OuterVolumeSpecName: "inventory") pod "7197a402-7e1b-450a-ba53-448bf843e6f6" (UID: "7197a402-7e1b-450a-ba53-448bf843e6f6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:49:14 crc kubenswrapper[4806]: I1127 10:49:14.617895 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7197a402-7e1b-450a-ba53-448bf843e6f6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7197a402-7e1b-450a-ba53-448bf843e6f6" (UID: "7197a402-7e1b-450a-ba53-448bf843e6f6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:49:14 crc kubenswrapper[4806]: I1127 10:49:14.691195 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7197a402-7e1b-450a-ba53-448bf843e6f6-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 10:49:14 crc kubenswrapper[4806]: I1127 10:49:14.691264 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngj9q\" (UniqueName: \"kubernetes.io/projected/7197a402-7e1b-450a-ba53-448bf843e6f6-kube-api-access-ngj9q\") on node \"crc\" DevicePath \"\"" Nov 27 10:49:14 crc kubenswrapper[4806]: I1127 10:49:14.691278 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7197a402-7e1b-450a-ba53-448bf843e6f6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.022854 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" event={"ID":"7197a402-7e1b-450a-ba53-448bf843e6f6","Type":"ContainerDied","Data":"e1342a8954d97f3a51d897eaf8d586818a5afe974f35419b010444155e0e12bf"} Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.023159 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1342a8954d97f3a51d897eaf8d586818a5afe974f35419b010444155e0e12bf" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.022948 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.100047 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq"] Nov 27 10:49:15 crc kubenswrapper[4806]: E1127 10:49:15.100549 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7197a402-7e1b-450a-ba53-448bf843e6f6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.100576 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7197a402-7e1b-450a-ba53-448bf843e6f6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.100862 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="7197a402-7e1b-450a-ba53-448bf843e6f6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.101673 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.117801 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.117894 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.117948 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.117906 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j9ghp" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.118811 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq"] Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.200644 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gjdvq\" (UID: \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.200776 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gjdvq\" (UID: \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.200827 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqqhd\" (UniqueName: \"kubernetes.io/projected/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-kube-api-access-cqqhd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gjdvq\" (UID: \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.302878 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gjdvq\" (UID: \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.303627 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gjdvq\" (UID: \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.304377 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqqhd\" (UniqueName: \"kubernetes.io/projected/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-kube-api-access-cqqhd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gjdvq\" (UID: \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.309402 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gjdvq\" (UID: \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.309612 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gjdvq\" (UID: \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.321024 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqqhd\" (UniqueName: \"kubernetes.io/projected/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-kube-api-access-cqqhd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-gjdvq\" (UID: \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" Nov 27 10:49:15 crc kubenswrapper[4806]: I1127 10:49:15.483792 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" Nov 27 10:49:16 crc kubenswrapper[4806]: W1127 10:49:16.029749 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33b9cdf1_651a_42c3_99b6_f62e1ef954bc.slice/crio-30f7e17ac9e4cd9c01ad62a49df823451a630b82af572293608019b1f91f68b4 WatchSource:0}: Error finding container 30f7e17ac9e4cd9c01ad62a49df823451a630b82af572293608019b1f91f68b4: Status 404 returned error can't find the container with id 30f7e17ac9e4cd9c01ad62a49df823451a630b82af572293608019b1f91f68b4 Nov 27 10:49:16 crc kubenswrapper[4806]: I1127 10:49:16.030427 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq"] Nov 27 10:49:17 crc kubenswrapper[4806]: I1127 10:49:17.061534 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" event={"ID":"33b9cdf1-651a-42c3-99b6-f62e1ef954bc","Type":"ContainerStarted","Data":"86b068a791caa39b3fb937042130091921bf3ecb25b0c60352bcc5160b1fb7b9"} Nov 27 10:49:17 crc kubenswrapper[4806]: I1127 10:49:17.062128 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" event={"ID":"33b9cdf1-651a-42c3-99b6-f62e1ef954bc","Type":"ContainerStarted","Data":"30f7e17ac9e4cd9c01ad62a49df823451a630b82af572293608019b1f91f68b4"} Nov 27 10:49:17 crc kubenswrapper[4806]: I1127 10:49:17.093182 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" podStartSLOduration=1.591446651 podStartE2EDuration="2.093154498s" podCreationTimestamp="2025-11-27 10:49:15 +0000 UTC" firstStartedPulling="2025-11-27 10:49:16.036076463 +0000 UTC m=+1660.622667227" lastFinishedPulling="2025-11-27 10:49:16.53778431 +0000 UTC m=+1661.124375074" observedRunningTime="2025-11-27 10:49:17.080163952 +0000 UTC m=+1661.666754716" watchObservedRunningTime="2025-11-27 10:49:17.093154498 +0000 UTC m=+1661.679745262" Nov 27 10:49:23 crc kubenswrapper[4806]: I1127 10:49:23.118114 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:49:23 crc kubenswrapper[4806]: E1127 10:49:23.119048 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:49:34 crc kubenswrapper[4806]: I1127 10:49:34.116699 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:49:34 crc kubenswrapper[4806]: E1127 10:49:34.117747 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:49:42 crc kubenswrapper[4806]: I1127 10:49:42.072304 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-4bc5-account-create-update-d58fg"] Nov 27 10:49:42 crc kubenswrapper[4806]: I1127 10:49:42.087649 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-28sdq"] Nov 27 10:49:42 crc kubenswrapper[4806]: I1127 10:49:42.096804 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-fjg6t"] Nov 27 10:49:42 crc kubenswrapper[4806]: I1127 10:49:42.106517 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-28sdq"] Nov 27 10:49:42 crc kubenswrapper[4806]: I1127 10:49:42.114894 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-8243-account-create-update-gwqg5"] Nov 27 10:49:42 crc kubenswrapper[4806]: I1127 10:49:42.128614 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7a0c344-9e24-43af-badd-b761a5005c56" path="/var/lib/kubelet/pods/e7a0c344-9e24-43af-badd-b761a5005c56/volumes" Nov 27 10:49:42 crc kubenswrapper[4806]: I1127 10:49:42.129592 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-8243-account-create-update-gwqg5"] Nov 27 10:49:42 crc kubenswrapper[4806]: I1127 10:49:42.134462 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-4bc5-account-create-update-d58fg"] Nov 27 10:49:42 crc kubenswrapper[4806]: I1127 10:49:42.141425 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-fjg6t"] Nov 27 10:49:43 crc kubenswrapper[4806]: I1127 10:49:43.033579 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-4fbch"] Nov 27 10:49:43 crc kubenswrapper[4806]: I1127 10:49:43.043313 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-4fbch"] Nov 27 10:49:43 crc kubenswrapper[4806]: I1127 10:49:43.053214 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-7795-account-create-update-wrpdm"] Nov 27 10:49:43 crc kubenswrapper[4806]: I1127 10:49:43.062100 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-7795-account-create-update-wrpdm"] Nov 27 10:49:44 crc kubenswrapper[4806]: I1127 10:49:44.146180 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="727574b8-48dd-4ca6-882f-e84934bf1240" path="/var/lib/kubelet/pods/727574b8-48dd-4ca6-882f-e84934bf1240/volumes" Nov 27 10:49:44 crc kubenswrapper[4806]: I1127 10:49:44.147294 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd4fca93-e550-4dff-a7d1-5efdd3d121b4" path="/var/lib/kubelet/pods/bd4fca93-e550-4dff-a7d1-5efdd3d121b4/volumes" Nov 27 10:49:44 crc kubenswrapper[4806]: I1127 10:49:44.148070 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3954c8e-9c73-4bb3-9db6-e64c3a86c480" path="/var/lib/kubelet/pods/d3954c8e-9c73-4bb3-9db6-e64c3a86c480/volumes" Nov 27 10:49:44 crc kubenswrapper[4806]: I1127 10:49:44.148661 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d436f2fa-66ed-4acb-970e-8b1116250622" path="/var/lib/kubelet/pods/d436f2fa-66ed-4acb-970e-8b1116250622/volumes" Nov 27 10:49:44 crc kubenswrapper[4806]: I1127 10:49:44.149711 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9041e87-4dd0-442d-a056-59d77e5134d0" path="/var/lib/kubelet/pods/f9041e87-4dd0-442d-a056-59d77e5134d0/volumes" Nov 27 10:49:47 crc kubenswrapper[4806]: I1127 10:49:47.117976 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:49:47 crc kubenswrapper[4806]: E1127 10:49:47.119828 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:49:53 crc kubenswrapper[4806]: I1127 10:49:53.469662 4806 scope.go:117] "RemoveContainer" containerID="f0444bce19c77f6a68d3d074f742f9ba2ff1217e4d92b5471c13275c0dc86ae1" Nov 27 10:49:53 crc kubenswrapper[4806]: I1127 10:49:53.514512 4806 scope.go:117] "RemoveContainer" containerID="529e41881d344faac4322ca1b807ae6ca76229cf4a751d348598a8dcec91c6e2" Nov 27 10:49:53 crc kubenswrapper[4806]: I1127 10:49:53.560507 4806 scope.go:117] "RemoveContainer" containerID="95cf89f7e44dd54684b95eb8164fa293fcc1e2c211c26814b25a739260a51543" Nov 27 10:49:53 crc kubenswrapper[4806]: I1127 10:49:53.610778 4806 scope.go:117] "RemoveContainer" containerID="f54b3e0a0fd3f20237c30e8fa9756b990de4f6a8be8489d3a1c2b7d823271841" Nov 27 10:49:53 crc kubenswrapper[4806]: I1127 10:49:53.674191 4806 scope.go:117] "RemoveContainer" containerID="e431d6ce993c73e597a8d790fb4f7749d7911b0e9afb0c85604899710d4a58cf" Nov 27 10:49:53 crc kubenswrapper[4806]: I1127 10:49:53.738343 4806 scope.go:117] "RemoveContainer" containerID="9717ca758a24a7bd89f9db2e4d21063a1025f77be0cd7c27b15e736c0864d702" Nov 27 10:49:53 crc kubenswrapper[4806]: I1127 10:49:53.759957 4806 scope.go:117] "RemoveContainer" containerID="3266f576d79cc8f500fe1c8806bde6016ef45abf6860873c153d950491afa787" Nov 27 10:49:59 crc kubenswrapper[4806]: I1127 10:49:59.118424 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:49:59 crc kubenswrapper[4806]: E1127 10:49:59.119642 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:50:04 crc kubenswrapper[4806]: I1127 10:50:04.613533 4806 generic.go:334] "Generic (PLEG): container finished" podID="33b9cdf1-651a-42c3-99b6-f62e1ef954bc" containerID="86b068a791caa39b3fb937042130091921bf3ecb25b0c60352bcc5160b1fb7b9" exitCode=0 Nov 27 10:50:04 crc kubenswrapper[4806]: I1127 10:50:04.613628 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" event={"ID":"33b9cdf1-651a-42c3-99b6-f62e1ef954bc","Type":"ContainerDied","Data":"86b068a791caa39b3fb937042130091921bf3ecb25b0c60352bcc5160b1fb7b9"} Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.171283 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.237563 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-inventory\") pod \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\" (UID: \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\") " Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.237939 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-ssh-key\") pod \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\" (UID: \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\") " Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.238188 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqqhd\" (UniqueName: \"kubernetes.io/projected/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-kube-api-access-cqqhd\") pod \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\" (UID: \"33b9cdf1-651a-42c3-99b6-f62e1ef954bc\") " Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.262102 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-kube-api-access-cqqhd" (OuterVolumeSpecName: "kube-api-access-cqqhd") pod "33b9cdf1-651a-42c3-99b6-f62e1ef954bc" (UID: "33b9cdf1-651a-42c3-99b6-f62e1ef954bc"). InnerVolumeSpecName "kube-api-access-cqqhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.270475 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "33b9cdf1-651a-42c3-99b6-f62e1ef954bc" (UID: "33b9cdf1-651a-42c3-99b6-f62e1ef954bc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.284464 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-inventory" (OuterVolumeSpecName: "inventory") pod "33b9cdf1-651a-42c3-99b6-f62e1ef954bc" (UID: "33b9cdf1-651a-42c3-99b6-f62e1ef954bc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.340682 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqqhd\" (UniqueName: \"kubernetes.io/projected/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-kube-api-access-cqqhd\") on node \"crc\" DevicePath \"\"" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.340732 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.340748 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/33b9cdf1-651a-42c3-99b6-f62e1ef954bc-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.639041 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" event={"ID":"33b9cdf1-651a-42c3-99b6-f62e1ef954bc","Type":"ContainerDied","Data":"30f7e17ac9e4cd9c01ad62a49df823451a630b82af572293608019b1f91f68b4"} Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.639093 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30f7e17ac9e4cd9c01ad62a49df823451a630b82af572293608019b1f91f68b4" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.639167 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-gjdvq" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.799420 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2"] Nov 27 10:50:06 crc kubenswrapper[4806]: E1127 10:50:06.799921 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b9cdf1-651a-42c3-99b6-f62e1ef954bc" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.799941 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b9cdf1-651a-42c3-99b6-f62e1ef954bc" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.800140 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b9cdf1-651a-42c3-99b6-f62e1ef954bc" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.800980 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.808470 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.809970 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.814178 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j9ghp" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.814954 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.851137 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2"] Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.975709 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e0dc02c-8ba2-485d-97eb-664f91b36046-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2\" (UID: \"1e0dc02c-8ba2-485d-97eb-664f91b36046\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.975864 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e0dc02c-8ba2-485d-97eb-664f91b36046-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2\" (UID: \"1e0dc02c-8ba2-485d-97eb-664f91b36046\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" Nov 27 10:50:06 crc kubenswrapper[4806]: I1127 10:50:06.975911 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28nw8\" (UniqueName: \"kubernetes.io/projected/1e0dc02c-8ba2-485d-97eb-664f91b36046-kube-api-access-28nw8\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2\" (UID: \"1e0dc02c-8ba2-485d-97eb-664f91b36046\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" Nov 27 10:50:07 crc kubenswrapper[4806]: I1127 10:50:07.077418 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e0dc02c-8ba2-485d-97eb-664f91b36046-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2\" (UID: \"1e0dc02c-8ba2-485d-97eb-664f91b36046\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" Nov 27 10:50:07 crc kubenswrapper[4806]: I1127 10:50:07.077492 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28nw8\" (UniqueName: \"kubernetes.io/projected/1e0dc02c-8ba2-485d-97eb-664f91b36046-kube-api-access-28nw8\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2\" (UID: \"1e0dc02c-8ba2-485d-97eb-664f91b36046\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" Nov 27 10:50:07 crc kubenswrapper[4806]: I1127 10:50:07.077565 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e0dc02c-8ba2-485d-97eb-664f91b36046-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2\" (UID: \"1e0dc02c-8ba2-485d-97eb-664f91b36046\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" Nov 27 10:50:07 crc kubenswrapper[4806]: I1127 10:50:07.084580 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e0dc02c-8ba2-485d-97eb-664f91b36046-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2\" (UID: \"1e0dc02c-8ba2-485d-97eb-664f91b36046\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" Nov 27 10:50:07 crc kubenswrapper[4806]: I1127 10:50:07.092341 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e0dc02c-8ba2-485d-97eb-664f91b36046-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2\" (UID: \"1e0dc02c-8ba2-485d-97eb-664f91b36046\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" Nov 27 10:50:07 crc kubenswrapper[4806]: I1127 10:50:07.107203 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28nw8\" (UniqueName: \"kubernetes.io/projected/1e0dc02c-8ba2-485d-97eb-664f91b36046-kube-api-access-28nw8\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2\" (UID: \"1e0dc02c-8ba2-485d-97eb-664f91b36046\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" Nov 27 10:50:07 crc kubenswrapper[4806]: I1127 10:50:07.150284 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" Nov 27 10:50:07 crc kubenswrapper[4806]: I1127 10:50:07.781161 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2"] Nov 27 10:50:08 crc kubenswrapper[4806]: I1127 10:50:08.658542 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" event={"ID":"1e0dc02c-8ba2-485d-97eb-664f91b36046","Type":"ContainerStarted","Data":"0d52525e91f4b1f5a4926c3f296cc54b4e8009d2f2628b467f3e89c5ec22f504"} Nov 27 10:50:08 crc kubenswrapper[4806]: I1127 10:50:08.659094 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" event={"ID":"1e0dc02c-8ba2-485d-97eb-664f91b36046","Type":"ContainerStarted","Data":"a29071d69af1f7b29a54a9204c12313a665b7ab8c65272037f5ff27c8f6e80e9"} Nov 27 10:50:08 crc kubenswrapper[4806]: I1127 10:50:08.683013 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" podStartSLOduration=2.158851893 podStartE2EDuration="2.682984617s" podCreationTimestamp="2025-11-27 10:50:06 +0000 UTC" firstStartedPulling="2025-11-27 10:50:07.793516657 +0000 UTC m=+1712.380107421" lastFinishedPulling="2025-11-27 10:50:08.317649361 +0000 UTC m=+1712.904240145" observedRunningTime="2025-11-27 10:50:08.676083441 +0000 UTC m=+1713.262674225" watchObservedRunningTime="2025-11-27 10:50:08.682984617 +0000 UTC m=+1713.269575401" Nov 27 10:50:11 crc kubenswrapper[4806]: I1127 10:50:11.117192 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:50:11 crc kubenswrapper[4806]: E1127 10:50:11.117880 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:50:12 crc kubenswrapper[4806]: I1127 10:50:12.063828 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bj4gg"] Nov 27 10:50:12 crc kubenswrapper[4806]: I1127 10:50:12.072296 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bj4gg"] Nov 27 10:50:12 crc kubenswrapper[4806]: I1127 10:50:12.129473 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7701f56-4e55-43dc-a526-3775fff2526c" path="/var/lib/kubelet/pods/d7701f56-4e55-43dc-a526-3775fff2526c/volumes" Nov 27 10:50:14 crc kubenswrapper[4806]: I1127 10:50:14.739485 4806 generic.go:334] "Generic (PLEG): container finished" podID="1e0dc02c-8ba2-485d-97eb-664f91b36046" containerID="0d52525e91f4b1f5a4926c3f296cc54b4e8009d2f2628b467f3e89c5ec22f504" exitCode=0 Nov 27 10:50:14 crc kubenswrapper[4806]: I1127 10:50:14.739527 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" event={"ID":"1e0dc02c-8ba2-485d-97eb-664f91b36046","Type":"ContainerDied","Data":"0d52525e91f4b1f5a4926c3f296cc54b4e8009d2f2628b467f3e89c5ec22f504"} Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.274199 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.403639 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e0dc02c-8ba2-485d-97eb-664f91b36046-ssh-key\") pod \"1e0dc02c-8ba2-485d-97eb-664f91b36046\" (UID: \"1e0dc02c-8ba2-485d-97eb-664f91b36046\") " Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.404047 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e0dc02c-8ba2-485d-97eb-664f91b36046-inventory\") pod \"1e0dc02c-8ba2-485d-97eb-664f91b36046\" (UID: \"1e0dc02c-8ba2-485d-97eb-664f91b36046\") " Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.404099 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28nw8\" (UniqueName: \"kubernetes.io/projected/1e0dc02c-8ba2-485d-97eb-664f91b36046-kube-api-access-28nw8\") pod \"1e0dc02c-8ba2-485d-97eb-664f91b36046\" (UID: \"1e0dc02c-8ba2-485d-97eb-664f91b36046\") " Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.417085 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e0dc02c-8ba2-485d-97eb-664f91b36046-kube-api-access-28nw8" (OuterVolumeSpecName: "kube-api-access-28nw8") pod "1e0dc02c-8ba2-485d-97eb-664f91b36046" (UID: "1e0dc02c-8ba2-485d-97eb-664f91b36046"). InnerVolumeSpecName "kube-api-access-28nw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.438720 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e0dc02c-8ba2-485d-97eb-664f91b36046-inventory" (OuterVolumeSpecName: "inventory") pod "1e0dc02c-8ba2-485d-97eb-664f91b36046" (UID: "1e0dc02c-8ba2-485d-97eb-664f91b36046"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.443174 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e0dc02c-8ba2-485d-97eb-664f91b36046-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1e0dc02c-8ba2-485d-97eb-664f91b36046" (UID: "1e0dc02c-8ba2-485d-97eb-664f91b36046"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.506395 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e0dc02c-8ba2-485d-97eb-664f91b36046-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.506443 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28nw8\" (UniqueName: \"kubernetes.io/projected/1e0dc02c-8ba2-485d-97eb-664f91b36046-kube-api-access-28nw8\") on node \"crc\" DevicePath \"\"" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.506459 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e0dc02c-8ba2-485d-97eb-664f91b36046-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.761373 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" event={"ID":"1e0dc02c-8ba2-485d-97eb-664f91b36046","Type":"ContainerDied","Data":"a29071d69af1f7b29a54a9204c12313a665b7ab8c65272037f5ff27c8f6e80e9"} Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.761429 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a29071d69af1f7b29a54a9204c12313a665b7ab8c65272037f5ff27c8f6e80e9" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.761444 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.862508 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4"] Nov 27 10:50:16 crc kubenswrapper[4806]: E1127 10:50:16.862996 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e0dc02c-8ba2-485d-97eb-664f91b36046" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.863012 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e0dc02c-8ba2-485d-97eb-664f91b36046" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.863266 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e0dc02c-8ba2-485d-97eb-664f91b36046" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.864027 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.866517 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.866817 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.866959 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j9ghp" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.867011 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 10:50:16 crc kubenswrapper[4806]: I1127 10:50:16.881787 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4"] Nov 27 10:50:17 crc kubenswrapper[4806]: I1127 10:50:17.020180 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqbsc\" (UniqueName: \"kubernetes.io/projected/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-kube-api-access-dqbsc\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-545s4\" (UID: \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" Nov 27 10:50:17 crc kubenswrapper[4806]: I1127 10:50:17.020383 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-545s4\" (UID: \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" Nov 27 10:50:17 crc kubenswrapper[4806]: I1127 10:50:17.020414 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-545s4\" (UID: \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" Nov 27 10:50:17 crc kubenswrapper[4806]: I1127 10:50:17.121507 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqbsc\" (UniqueName: \"kubernetes.io/projected/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-kube-api-access-dqbsc\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-545s4\" (UID: \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" Nov 27 10:50:17 crc kubenswrapper[4806]: I1127 10:50:17.121558 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-545s4\" (UID: \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" Nov 27 10:50:17 crc kubenswrapper[4806]: I1127 10:50:17.121584 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-545s4\" (UID: \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" Nov 27 10:50:17 crc kubenswrapper[4806]: I1127 10:50:17.127293 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-545s4\" (UID: \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" Nov 27 10:50:17 crc kubenswrapper[4806]: I1127 10:50:17.130822 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-545s4\" (UID: \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" Nov 27 10:50:17 crc kubenswrapper[4806]: I1127 10:50:17.143456 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqbsc\" (UniqueName: \"kubernetes.io/projected/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-kube-api-access-dqbsc\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-545s4\" (UID: \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" Nov 27 10:50:17 crc kubenswrapper[4806]: I1127 10:50:17.189438 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" Nov 27 10:50:18 crc kubenswrapper[4806]: I1127 10:50:18.164020 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4"] Nov 27 10:50:18 crc kubenswrapper[4806]: I1127 10:50:18.789070 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" event={"ID":"2699b50a-1d3a-4e3b-9b01-b67f06d04f77","Type":"ContainerStarted","Data":"e57b402f30313a06c33251048e248af6b684c5c6ada224a41e2e208ec53b45cb"} Nov 27 10:50:19 crc kubenswrapper[4806]: I1127 10:50:19.802651 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" event={"ID":"2699b50a-1d3a-4e3b-9b01-b67f06d04f77","Type":"ContainerStarted","Data":"5549d4dfc926a55ee81692f6672443d4177a94386c20b5f8c3278c2bb115eeb4"} Nov 27 10:50:19 crc kubenswrapper[4806]: I1127 10:50:19.829659 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" podStartSLOduration=3.156814492 podStartE2EDuration="3.829634472s" podCreationTimestamp="2025-11-27 10:50:16 +0000 UTC" firstStartedPulling="2025-11-27 10:50:18.139828849 +0000 UTC m=+1722.726419623" lastFinishedPulling="2025-11-27 10:50:18.812648839 +0000 UTC m=+1723.399239603" observedRunningTime="2025-11-27 10:50:19.823481006 +0000 UTC m=+1724.410071770" watchObservedRunningTime="2025-11-27 10:50:19.829634472 +0000 UTC m=+1724.416225236" Nov 27 10:50:22 crc kubenswrapper[4806]: I1127 10:50:22.118038 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:50:22 crc kubenswrapper[4806]: E1127 10:50:22.118908 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:50:37 crc kubenswrapper[4806]: I1127 10:50:37.065725 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-zhv9q"] Nov 27 10:50:37 crc kubenswrapper[4806]: I1127 10:50:37.078521 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6j7m6"] Nov 27 10:50:37 crc kubenswrapper[4806]: I1127 10:50:37.098861 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6j7m6"] Nov 27 10:50:37 crc kubenswrapper[4806]: I1127 10:50:37.126699 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:50:37 crc kubenswrapper[4806]: E1127 10:50:37.132357 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:50:37 crc kubenswrapper[4806]: I1127 10:50:37.141869 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-zhv9q"] Nov 27 10:50:38 crc kubenswrapper[4806]: I1127 10:50:38.130993 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b0f7898-ee70-46b9-a405-54419ec98b47" path="/var/lib/kubelet/pods/2b0f7898-ee70-46b9-a405-54419ec98b47/volumes" Nov 27 10:50:38 crc kubenswrapper[4806]: I1127 10:50:38.132871 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b15a49b-4c6b-4574-a229-25ea19ea2c0f" path="/var/lib/kubelet/pods/9b15a49b-4c6b-4574-a229-25ea19ea2c0f/volumes" Nov 27 10:50:42 crc kubenswrapper[4806]: I1127 10:50:42.455968 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gph9d"] Nov 27 10:50:42 crc kubenswrapper[4806]: I1127 10:50:42.458456 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:50:42 crc kubenswrapper[4806]: I1127 10:50:42.489170 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gph9d"] Nov 27 10:50:42 crc kubenswrapper[4806]: I1127 10:50:42.537433 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c83fe3b-c047-40e3-9a06-8460f178a7fb-utilities\") pod \"redhat-operators-gph9d\" (UID: \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\") " pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:50:42 crc kubenswrapper[4806]: I1127 10:50:42.537732 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c83fe3b-c047-40e3-9a06-8460f178a7fb-catalog-content\") pod \"redhat-operators-gph9d\" (UID: \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\") " pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:50:42 crc kubenswrapper[4806]: I1127 10:50:42.538087 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtbvb\" (UniqueName: \"kubernetes.io/projected/5c83fe3b-c047-40e3-9a06-8460f178a7fb-kube-api-access-qtbvb\") pod \"redhat-operators-gph9d\" (UID: \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\") " pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:50:42 crc kubenswrapper[4806]: I1127 10:50:42.640851 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c83fe3b-c047-40e3-9a06-8460f178a7fb-catalog-content\") pod \"redhat-operators-gph9d\" (UID: \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\") " pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:50:42 crc kubenswrapper[4806]: I1127 10:50:42.640980 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtbvb\" (UniqueName: \"kubernetes.io/projected/5c83fe3b-c047-40e3-9a06-8460f178a7fb-kube-api-access-qtbvb\") pod \"redhat-operators-gph9d\" (UID: \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\") " pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:50:42 crc kubenswrapper[4806]: I1127 10:50:42.641055 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c83fe3b-c047-40e3-9a06-8460f178a7fb-utilities\") pod \"redhat-operators-gph9d\" (UID: \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\") " pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:50:42 crc kubenswrapper[4806]: I1127 10:50:42.641524 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c83fe3b-c047-40e3-9a06-8460f178a7fb-catalog-content\") pod \"redhat-operators-gph9d\" (UID: \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\") " pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:50:42 crc kubenswrapper[4806]: I1127 10:50:42.641699 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c83fe3b-c047-40e3-9a06-8460f178a7fb-utilities\") pod \"redhat-operators-gph9d\" (UID: \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\") " pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:50:42 crc kubenswrapper[4806]: I1127 10:50:42.667295 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtbvb\" (UniqueName: \"kubernetes.io/projected/5c83fe3b-c047-40e3-9a06-8460f178a7fb-kube-api-access-qtbvb\") pod \"redhat-operators-gph9d\" (UID: \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\") " pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:50:42 crc kubenswrapper[4806]: I1127 10:50:42.793675 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:50:43 crc kubenswrapper[4806]: I1127 10:50:43.344215 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gph9d"] Nov 27 10:50:44 crc kubenswrapper[4806]: I1127 10:50:44.239040 4806 generic.go:334] "Generic (PLEG): container finished" podID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" containerID="dd6bd6d14332225e5b33a92da4b11ebcd1696c111224be0aab5fcc528608cce8" exitCode=0 Nov 27 10:50:44 crc kubenswrapper[4806]: I1127 10:50:44.239404 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gph9d" event={"ID":"5c83fe3b-c047-40e3-9a06-8460f178a7fb","Type":"ContainerDied","Data":"dd6bd6d14332225e5b33a92da4b11ebcd1696c111224be0aab5fcc528608cce8"} Nov 27 10:50:44 crc kubenswrapper[4806]: I1127 10:50:44.239503 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gph9d" event={"ID":"5c83fe3b-c047-40e3-9a06-8460f178a7fb","Type":"ContainerStarted","Data":"3268bbc96ac9c996267416fd5a4147716a0edbebb002bbaa3dc07aa52c1273ad"} Nov 27 10:50:46 crc kubenswrapper[4806]: I1127 10:50:46.263172 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gph9d" event={"ID":"5c83fe3b-c047-40e3-9a06-8460f178a7fb","Type":"ContainerStarted","Data":"b3a781c1027ef27b0bf433520745f2c101d1d4fae0f6b74aeb0a6833fd7bfbce"} Nov 27 10:50:50 crc kubenswrapper[4806]: I1127 10:50:50.117540 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:50:50 crc kubenswrapper[4806]: E1127 10:50:50.119046 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:50:51 crc kubenswrapper[4806]: I1127 10:50:51.307495 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gph9d" event={"ID":"5c83fe3b-c047-40e3-9a06-8460f178a7fb","Type":"ContainerDied","Data":"b3a781c1027ef27b0bf433520745f2c101d1d4fae0f6b74aeb0a6833fd7bfbce"} Nov 27 10:50:51 crc kubenswrapper[4806]: I1127 10:50:51.307541 4806 generic.go:334] "Generic (PLEG): container finished" podID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" containerID="b3a781c1027ef27b0bf433520745f2c101d1d4fae0f6b74aeb0a6833fd7bfbce" exitCode=0 Nov 27 10:50:52 crc kubenswrapper[4806]: I1127 10:50:52.322292 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gph9d" event={"ID":"5c83fe3b-c047-40e3-9a06-8460f178a7fb","Type":"ContainerStarted","Data":"5d997471906af39f5df68db360eb7ef908dff99b7a528a2ba9694bf471f47b12"} Nov 27 10:50:52 crc kubenswrapper[4806]: I1127 10:50:52.358916 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gph9d" podStartSLOduration=2.665389321 podStartE2EDuration="10.358882494s" podCreationTimestamp="2025-11-27 10:50:42 +0000 UTC" firstStartedPulling="2025-11-27 10:50:44.24579256 +0000 UTC m=+1748.832383324" lastFinishedPulling="2025-11-27 10:50:51.939285703 +0000 UTC m=+1756.525876497" observedRunningTime="2025-11-27 10:50:52.357173718 +0000 UTC m=+1756.943764492" watchObservedRunningTime="2025-11-27 10:50:52.358882494 +0000 UTC m=+1756.945473258" Nov 27 10:50:52 crc kubenswrapper[4806]: I1127 10:50:52.794462 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:50:52 crc kubenswrapper[4806]: I1127 10:50:52.794513 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:50:53 crc kubenswrapper[4806]: I1127 10:50:53.842991 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gph9d" podUID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" containerName="registry-server" probeResult="failure" output=< Nov 27 10:50:53 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 10:50:53 crc kubenswrapper[4806]: > Nov 27 10:50:53 crc kubenswrapper[4806]: I1127 10:50:53.918414 4806 scope.go:117] "RemoveContainer" containerID="f5683037cf95e466d7e550063cda0093b402b439d511b553b012f97b5f32bc23" Nov 27 10:50:53 crc kubenswrapper[4806]: I1127 10:50:53.980560 4806 scope.go:117] "RemoveContainer" containerID="af60cd81fcfdc97a1eac9448be597d2b6886c711e903e6e66d33faec24bd354d" Nov 27 10:50:54 crc kubenswrapper[4806]: I1127 10:50:54.022390 4806 scope.go:117] "RemoveContainer" containerID="aa917caab719a500e5bc3e742e1ae89c28086ef0cddd38339c19116db015a166" Nov 27 10:51:03 crc kubenswrapper[4806]: I1127 10:51:03.866639 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gph9d" podUID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" containerName="registry-server" probeResult="failure" output=< Nov 27 10:51:03 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 10:51:03 crc kubenswrapper[4806]: > Nov 27 10:51:05 crc kubenswrapper[4806]: I1127 10:51:05.117104 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:51:05 crc kubenswrapper[4806]: E1127 10:51:05.117965 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:51:12 crc kubenswrapper[4806]: I1127 10:51:12.862819 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:51:12 crc kubenswrapper[4806]: I1127 10:51:12.924077 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:51:13 crc kubenswrapper[4806]: I1127 10:51:13.106108 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gph9d"] Nov 27 10:51:14 crc kubenswrapper[4806]: I1127 10:51:14.557686 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gph9d" podUID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" containerName="registry-server" containerID="cri-o://5d997471906af39f5df68db360eb7ef908dff99b7a528a2ba9694bf471f47b12" gracePeriod=2 Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.029787 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.169649 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c83fe3b-c047-40e3-9a06-8460f178a7fb-catalog-content\") pod \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\" (UID: \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\") " Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.169854 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtbvb\" (UniqueName: \"kubernetes.io/projected/5c83fe3b-c047-40e3-9a06-8460f178a7fb-kube-api-access-qtbvb\") pod \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\" (UID: \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\") " Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.170010 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c83fe3b-c047-40e3-9a06-8460f178a7fb-utilities\") pod \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\" (UID: \"5c83fe3b-c047-40e3-9a06-8460f178a7fb\") " Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.170823 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c83fe3b-c047-40e3-9a06-8460f178a7fb-utilities" (OuterVolumeSpecName: "utilities") pod "5c83fe3b-c047-40e3-9a06-8460f178a7fb" (UID: "5c83fe3b-c047-40e3-9a06-8460f178a7fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.178467 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c83fe3b-c047-40e3-9a06-8460f178a7fb-kube-api-access-qtbvb" (OuterVolumeSpecName: "kube-api-access-qtbvb") pod "5c83fe3b-c047-40e3-9a06-8460f178a7fb" (UID: "5c83fe3b-c047-40e3-9a06-8460f178a7fb"). InnerVolumeSpecName "kube-api-access-qtbvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.273483 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtbvb\" (UniqueName: \"kubernetes.io/projected/5c83fe3b-c047-40e3-9a06-8460f178a7fb-kube-api-access-qtbvb\") on node \"crc\" DevicePath \"\"" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.273521 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c83fe3b-c047-40e3-9a06-8460f178a7fb-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.279521 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c83fe3b-c047-40e3-9a06-8460f178a7fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c83fe3b-c047-40e3-9a06-8460f178a7fb" (UID: "5c83fe3b-c047-40e3-9a06-8460f178a7fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.375350 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c83fe3b-c047-40e3-9a06-8460f178a7fb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.569799 4806 generic.go:334] "Generic (PLEG): container finished" podID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" containerID="5d997471906af39f5df68db360eb7ef908dff99b7a528a2ba9694bf471f47b12" exitCode=0 Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.569861 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gph9d" event={"ID":"5c83fe3b-c047-40e3-9a06-8460f178a7fb","Type":"ContainerDied","Data":"5d997471906af39f5df68db360eb7ef908dff99b7a528a2ba9694bf471f47b12"} Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.569899 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gph9d" event={"ID":"5c83fe3b-c047-40e3-9a06-8460f178a7fb","Type":"ContainerDied","Data":"3268bbc96ac9c996267416fd5a4147716a0edbebb002bbaa3dc07aa52c1273ad"} Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.569920 4806 scope.go:117] "RemoveContainer" containerID="5d997471906af39f5df68db360eb7ef908dff99b7a528a2ba9694bf471f47b12" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.569997 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gph9d" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.595707 4806 scope.go:117] "RemoveContainer" containerID="b3a781c1027ef27b0bf433520745f2c101d1d4fae0f6b74aeb0a6833fd7bfbce" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.612163 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gph9d"] Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.620798 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gph9d"] Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.630717 4806 scope.go:117] "RemoveContainer" containerID="dd6bd6d14332225e5b33a92da4b11ebcd1696c111224be0aab5fcc528608cce8" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.689206 4806 scope.go:117] "RemoveContainer" containerID="5d997471906af39f5df68db360eb7ef908dff99b7a528a2ba9694bf471f47b12" Nov 27 10:51:15 crc kubenswrapper[4806]: E1127 10:51:15.692509 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d997471906af39f5df68db360eb7ef908dff99b7a528a2ba9694bf471f47b12\": container with ID starting with 5d997471906af39f5df68db360eb7ef908dff99b7a528a2ba9694bf471f47b12 not found: ID does not exist" containerID="5d997471906af39f5df68db360eb7ef908dff99b7a528a2ba9694bf471f47b12" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.692571 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d997471906af39f5df68db360eb7ef908dff99b7a528a2ba9694bf471f47b12"} err="failed to get container status \"5d997471906af39f5df68db360eb7ef908dff99b7a528a2ba9694bf471f47b12\": rpc error: code = NotFound desc = could not find container \"5d997471906af39f5df68db360eb7ef908dff99b7a528a2ba9694bf471f47b12\": container with ID starting with 5d997471906af39f5df68db360eb7ef908dff99b7a528a2ba9694bf471f47b12 not found: ID does not exist" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.692604 4806 scope.go:117] "RemoveContainer" containerID="b3a781c1027ef27b0bf433520745f2c101d1d4fae0f6b74aeb0a6833fd7bfbce" Nov 27 10:51:15 crc kubenswrapper[4806]: E1127 10:51:15.693103 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3a781c1027ef27b0bf433520745f2c101d1d4fae0f6b74aeb0a6833fd7bfbce\": container with ID starting with b3a781c1027ef27b0bf433520745f2c101d1d4fae0f6b74aeb0a6833fd7bfbce not found: ID does not exist" containerID="b3a781c1027ef27b0bf433520745f2c101d1d4fae0f6b74aeb0a6833fd7bfbce" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.693177 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3a781c1027ef27b0bf433520745f2c101d1d4fae0f6b74aeb0a6833fd7bfbce"} err="failed to get container status \"b3a781c1027ef27b0bf433520745f2c101d1d4fae0f6b74aeb0a6833fd7bfbce\": rpc error: code = NotFound desc = could not find container \"b3a781c1027ef27b0bf433520745f2c101d1d4fae0f6b74aeb0a6833fd7bfbce\": container with ID starting with b3a781c1027ef27b0bf433520745f2c101d1d4fae0f6b74aeb0a6833fd7bfbce not found: ID does not exist" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.693215 4806 scope.go:117] "RemoveContainer" containerID="dd6bd6d14332225e5b33a92da4b11ebcd1696c111224be0aab5fcc528608cce8" Nov 27 10:51:15 crc kubenswrapper[4806]: E1127 10:51:15.693728 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd6bd6d14332225e5b33a92da4b11ebcd1696c111224be0aab5fcc528608cce8\": container with ID starting with dd6bd6d14332225e5b33a92da4b11ebcd1696c111224be0aab5fcc528608cce8 not found: ID does not exist" containerID="dd6bd6d14332225e5b33a92da4b11ebcd1696c111224be0aab5fcc528608cce8" Nov 27 10:51:15 crc kubenswrapper[4806]: I1127 10:51:15.693857 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd6bd6d14332225e5b33a92da4b11ebcd1696c111224be0aab5fcc528608cce8"} err="failed to get container status \"dd6bd6d14332225e5b33a92da4b11ebcd1696c111224be0aab5fcc528608cce8\": rpc error: code = NotFound desc = could not find container \"dd6bd6d14332225e5b33a92da4b11ebcd1696c111224be0aab5fcc528608cce8\": container with ID starting with dd6bd6d14332225e5b33a92da4b11ebcd1696c111224be0aab5fcc528608cce8 not found: ID does not exist" Nov 27 10:51:16 crc kubenswrapper[4806]: I1127 10:51:16.128572 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:51:16 crc kubenswrapper[4806]: I1127 10:51:16.143415 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" path="/var/lib/kubelet/pods/5c83fe3b-c047-40e3-9a06-8460f178a7fb/volumes" Nov 27 10:51:17 crc kubenswrapper[4806]: I1127 10:51:17.602903 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"4499112e23b5522078d051c09243e588431a647460adb8b3e111de397927146b"} Nov 27 10:51:23 crc kubenswrapper[4806]: I1127 10:51:23.048913 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-zjnmw"] Nov 27 10:51:23 crc kubenswrapper[4806]: I1127 10:51:23.057193 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-zjnmw"] Nov 27 10:51:24 crc kubenswrapper[4806]: I1127 10:51:24.130309 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cce234e3-7d48-4640-8f9d-0e7fac75f77e" path="/var/lib/kubelet/pods/cce234e3-7d48-4640-8f9d-0e7fac75f77e/volumes" Nov 27 10:51:28 crc kubenswrapper[4806]: I1127 10:51:28.827045 4806 generic.go:334] "Generic (PLEG): container finished" podID="2699b50a-1d3a-4e3b-9b01-b67f06d04f77" containerID="5549d4dfc926a55ee81692f6672443d4177a94386c20b5f8c3278c2bb115eeb4" exitCode=0 Nov 27 10:51:28 crc kubenswrapper[4806]: I1127 10:51:28.827165 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" event={"ID":"2699b50a-1d3a-4e3b-9b01-b67f06d04f77","Type":"ContainerDied","Data":"5549d4dfc926a55ee81692f6672443d4177a94386c20b5f8c3278c2bb115eeb4"} Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.371317 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.494223 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-inventory\") pod \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\" (UID: \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\") " Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.494321 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqbsc\" (UniqueName: \"kubernetes.io/projected/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-kube-api-access-dqbsc\") pod \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\" (UID: \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\") " Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.494482 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-ssh-key\") pod \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\" (UID: \"2699b50a-1d3a-4e3b-9b01-b67f06d04f77\") " Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.505622 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-kube-api-access-dqbsc" (OuterVolumeSpecName: "kube-api-access-dqbsc") pod "2699b50a-1d3a-4e3b-9b01-b67f06d04f77" (UID: "2699b50a-1d3a-4e3b-9b01-b67f06d04f77"). InnerVolumeSpecName "kube-api-access-dqbsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.527912 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-inventory" (OuterVolumeSpecName: "inventory") pod "2699b50a-1d3a-4e3b-9b01-b67f06d04f77" (UID: "2699b50a-1d3a-4e3b-9b01-b67f06d04f77"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.531006 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2699b50a-1d3a-4e3b-9b01-b67f06d04f77" (UID: "2699b50a-1d3a-4e3b-9b01-b67f06d04f77"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.597560 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.598330 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.598349 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqbsc\" (UniqueName: \"kubernetes.io/projected/2699b50a-1d3a-4e3b-9b01-b67f06d04f77-kube-api-access-dqbsc\") on node \"crc\" DevicePath \"\"" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.850392 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" event={"ID":"2699b50a-1d3a-4e3b-9b01-b67f06d04f77","Type":"ContainerDied","Data":"e57b402f30313a06c33251048e248af6b684c5c6ada224a41e2e208ec53b45cb"} Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.850449 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e57b402f30313a06c33251048e248af6b684c5c6ada224a41e2e208ec53b45cb" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.850485 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-545s4" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.949078 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-djqnt"] Nov 27 10:51:30 crc kubenswrapper[4806]: E1127 10:51:30.949548 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" containerName="extract-content" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.949570 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" containerName="extract-content" Nov 27 10:51:30 crc kubenswrapper[4806]: E1127 10:51:30.949583 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2699b50a-1d3a-4e3b-9b01-b67f06d04f77" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.949602 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2699b50a-1d3a-4e3b-9b01-b67f06d04f77" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 27 10:51:30 crc kubenswrapper[4806]: E1127 10:51:30.949659 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" containerName="registry-server" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.949666 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" containerName="registry-server" Nov 27 10:51:30 crc kubenswrapper[4806]: E1127 10:51:30.949682 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" containerName="extract-utilities" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.949689 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" containerName="extract-utilities" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.949876 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c83fe3b-c047-40e3-9a06-8460f178a7fb" containerName="registry-server" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.949898 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="2699b50a-1d3a-4e3b-9b01-b67f06d04f77" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.950677 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.957791 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.958031 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.958163 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j9ghp" Nov 27 10:51:30 crc kubenswrapper[4806]: I1127 10:51:30.982681 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 10:51:31 crc kubenswrapper[4806]: I1127 10:51:31.005485 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-djqnt"] Nov 27 10:51:31 crc kubenswrapper[4806]: I1127 10:51:31.110653 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/029a4e3d-528c-47a0-b34d-42f296055ed0-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-djqnt\" (UID: \"029a4e3d-528c-47a0-b34d-42f296055ed0\") " pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" Nov 27 10:51:31 crc kubenswrapper[4806]: I1127 10:51:31.111087 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkvqx\" (UniqueName: \"kubernetes.io/projected/029a4e3d-528c-47a0-b34d-42f296055ed0-kube-api-access-pkvqx\") pod \"ssh-known-hosts-edpm-deployment-djqnt\" (UID: \"029a4e3d-528c-47a0-b34d-42f296055ed0\") " pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" Nov 27 10:51:31 crc kubenswrapper[4806]: I1127 10:51:31.111158 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/029a4e3d-528c-47a0-b34d-42f296055ed0-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-djqnt\" (UID: \"029a4e3d-528c-47a0-b34d-42f296055ed0\") " pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" Nov 27 10:51:31 crc kubenswrapper[4806]: I1127 10:51:31.214312 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/029a4e3d-528c-47a0-b34d-42f296055ed0-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-djqnt\" (UID: \"029a4e3d-528c-47a0-b34d-42f296055ed0\") " pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" Nov 27 10:51:31 crc kubenswrapper[4806]: I1127 10:51:31.214361 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkvqx\" (UniqueName: \"kubernetes.io/projected/029a4e3d-528c-47a0-b34d-42f296055ed0-kube-api-access-pkvqx\") pod \"ssh-known-hosts-edpm-deployment-djqnt\" (UID: \"029a4e3d-528c-47a0-b34d-42f296055ed0\") " pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" Nov 27 10:51:31 crc kubenswrapper[4806]: I1127 10:51:31.214416 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/029a4e3d-528c-47a0-b34d-42f296055ed0-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-djqnt\" (UID: \"029a4e3d-528c-47a0-b34d-42f296055ed0\") " pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" Nov 27 10:51:31 crc kubenswrapper[4806]: I1127 10:51:31.221289 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/029a4e3d-528c-47a0-b34d-42f296055ed0-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-djqnt\" (UID: \"029a4e3d-528c-47a0-b34d-42f296055ed0\") " pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" Nov 27 10:51:31 crc kubenswrapper[4806]: I1127 10:51:31.222910 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/029a4e3d-528c-47a0-b34d-42f296055ed0-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-djqnt\" (UID: \"029a4e3d-528c-47a0-b34d-42f296055ed0\") " pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" Nov 27 10:51:31 crc kubenswrapper[4806]: I1127 10:51:31.235936 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkvqx\" (UniqueName: \"kubernetes.io/projected/029a4e3d-528c-47a0-b34d-42f296055ed0-kube-api-access-pkvqx\") pod \"ssh-known-hosts-edpm-deployment-djqnt\" (UID: \"029a4e3d-528c-47a0-b34d-42f296055ed0\") " pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" Nov 27 10:51:31 crc kubenswrapper[4806]: I1127 10:51:31.275187 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" Nov 27 10:51:32 crc kubenswrapper[4806]: I1127 10:51:31.894813 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-djqnt"] Nov 27 10:51:32 crc kubenswrapper[4806]: I1127 10:51:32.872346 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" event={"ID":"029a4e3d-528c-47a0-b34d-42f296055ed0","Type":"ContainerStarted","Data":"49b4c0d170010a33372ec38b1f46b97230c404b744e292f092937da386891910"} Nov 27 10:51:32 crc kubenswrapper[4806]: I1127 10:51:32.872851 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" event={"ID":"029a4e3d-528c-47a0-b34d-42f296055ed0","Type":"ContainerStarted","Data":"aa37560a56b3e2f6e1dffd6d95712433740dea7cedef1590757861957fcd02b8"} Nov 27 10:51:32 crc kubenswrapper[4806]: I1127 10:51:32.897450 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" podStartSLOduration=2.24030228 podStartE2EDuration="2.897415936s" podCreationTimestamp="2025-11-27 10:51:30 +0000 UTC" firstStartedPulling="2025-11-27 10:51:31.916340201 +0000 UTC m=+1796.502930965" lastFinishedPulling="2025-11-27 10:51:32.573453857 +0000 UTC m=+1797.160044621" observedRunningTime="2025-11-27 10:51:32.887808816 +0000 UTC m=+1797.474399580" watchObservedRunningTime="2025-11-27 10:51:32.897415936 +0000 UTC m=+1797.484006700" Nov 27 10:51:42 crc kubenswrapper[4806]: I1127 10:51:42.978660 4806 generic.go:334] "Generic (PLEG): container finished" podID="029a4e3d-528c-47a0-b34d-42f296055ed0" containerID="49b4c0d170010a33372ec38b1f46b97230c404b744e292f092937da386891910" exitCode=0 Nov 27 10:51:42 crc kubenswrapper[4806]: I1127 10:51:42.978722 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" event={"ID":"029a4e3d-528c-47a0-b34d-42f296055ed0","Type":"ContainerDied","Data":"49b4c0d170010a33372ec38b1f46b97230c404b744e292f092937da386891910"} Nov 27 10:51:44 crc kubenswrapper[4806]: I1127 10:51:44.450331 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" Nov 27 10:51:44 crc kubenswrapper[4806]: I1127 10:51:44.615675 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/029a4e3d-528c-47a0-b34d-42f296055ed0-ssh-key-openstack-edpm-ipam\") pod \"029a4e3d-528c-47a0-b34d-42f296055ed0\" (UID: \"029a4e3d-528c-47a0-b34d-42f296055ed0\") " Nov 27 10:51:44 crc kubenswrapper[4806]: I1127 10:51:44.615803 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/029a4e3d-528c-47a0-b34d-42f296055ed0-inventory-0\") pod \"029a4e3d-528c-47a0-b34d-42f296055ed0\" (UID: \"029a4e3d-528c-47a0-b34d-42f296055ed0\") " Nov 27 10:51:44 crc kubenswrapper[4806]: I1127 10:51:44.615880 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkvqx\" (UniqueName: \"kubernetes.io/projected/029a4e3d-528c-47a0-b34d-42f296055ed0-kube-api-access-pkvqx\") pod \"029a4e3d-528c-47a0-b34d-42f296055ed0\" (UID: \"029a4e3d-528c-47a0-b34d-42f296055ed0\") " Nov 27 10:51:44 crc kubenswrapper[4806]: I1127 10:51:44.628578 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/029a4e3d-528c-47a0-b34d-42f296055ed0-kube-api-access-pkvqx" (OuterVolumeSpecName: "kube-api-access-pkvqx") pod "029a4e3d-528c-47a0-b34d-42f296055ed0" (UID: "029a4e3d-528c-47a0-b34d-42f296055ed0"). InnerVolumeSpecName "kube-api-access-pkvqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:51:44 crc kubenswrapper[4806]: I1127 10:51:44.647870 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/029a4e3d-528c-47a0-b34d-42f296055ed0-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "029a4e3d-528c-47a0-b34d-42f296055ed0" (UID: "029a4e3d-528c-47a0-b34d-42f296055ed0"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:51:44 crc kubenswrapper[4806]: I1127 10:51:44.649477 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/029a4e3d-528c-47a0-b34d-42f296055ed0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "029a4e3d-528c-47a0-b34d-42f296055ed0" (UID: "029a4e3d-528c-47a0-b34d-42f296055ed0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:51:44 crc kubenswrapper[4806]: I1127 10:51:44.718726 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/029a4e3d-528c-47a0-b34d-42f296055ed0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 27 10:51:44 crc kubenswrapper[4806]: I1127 10:51:44.718783 4806 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/029a4e3d-528c-47a0-b34d-42f296055ed0-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 27 10:51:44 crc kubenswrapper[4806]: I1127 10:51:44.718803 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkvqx\" (UniqueName: \"kubernetes.io/projected/029a4e3d-528c-47a0-b34d-42f296055ed0-kube-api-access-pkvqx\") on node \"crc\" DevicePath \"\"" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.008327 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" event={"ID":"029a4e3d-528c-47a0-b34d-42f296055ed0","Type":"ContainerDied","Data":"aa37560a56b3e2f6e1dffd6d95712433740dea7cedef1590757861957fcd02b8"} Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.008761 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa37560a56b3e2f6e1dffd6d95712433740dea7cedef1590757861957fcd02b8" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.008405 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-djqnt" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.126850 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb"] Nov 27 10:51:45 crc kubenswrapper[4806]: E1127 10:51:45.127470 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="029a4e3d-528c-47a0-b34d-42f296055ed0" containerName="ssh-known-hosts-edpm-deployment" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.127494 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="029a4e3d-528c-47a0-b34d-42f296055ed0" containerName="ssh-known-hosts-edpm-deployment" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.128000 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="029a4e3d-528c-47a0-b34d-42f296055ed0" containerName="ssh-known-hosts-edpm-deployment" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.129055 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.132920 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.133122 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j9ghp" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.133255 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.133474 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.157824 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb"] Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.228903 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c358c0e7-2566-4f18-8a4a-2af284af91f6-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qvprb\" (UID: \"c358c0e7-2566-4f18-8a4a-2af284af91f6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.229242 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c358c0e7-2566-4f18-8a4a-2af284af91f6-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qvprb\" (UID: \"c358c0e7-2566-4f18-8a4a-2af284af91f6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.229314 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxr2q\" (UniqueName: \"kubernetes.io/projected/c358c0e7-2566-4f18-8a4a-2af284af91f6-kube-api-access-mxr2q\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qvprb\" (UID: \"c358c0e7-2566-4f18-8a4a-2af284af91f6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.333703 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c358c0e7-2566-4f18-8a4a-2af284af91f6-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qvprb\" (UID: \"c358c0e7-2566-4f18-8a4a-2af284af91f6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.333808 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c358c0e7-2566-4f18-8a4a-2af284af91f6-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qvprb\" (UID: \"c358c0e7-2566-4f18-8a4a-2af284af91f6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.333829 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxr2q\" (UniqueName: \"kubernetes.io/projected/c358c0e7-2566-4f18-8a4a-2af284af91f6-kube-api-access-mxr2q\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qvprb\" (UID: \"c358c0e7-2566-4f18-8a4a-2af284af91f6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.338852 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c358c0e7-2566-4f18-8a4a-2af284af91f6-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qvprb\" (UID: \"c358c0e7-2566-4f18-8a4a-2af284af91f6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.342392 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c358c0e7-2566-4f18-8a4a-2af284af91f6-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qvprb\" (UID: \"c358c0e7-2566-4f18-8a4a-2af284af91f6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.354438 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxr2q\" (UniqueName: \"kubernetes.io/projected/c358c0e7-2566-4f18-8a4a-2af284af91f6-kube-api-access-mxr2q\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qvprb\" (UID: \"c358c0e7-2566-4f18-8a4a-2af284af91f6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" Nov 27 10:51:45 crc kubenswrapper[4806]: I1127 10:51:45.452564 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" Nov 27 10:51:46 crc kubenswrapper[4806]: I1127 10:51:46.029459 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb"] Nov 27 10:51:46 crc kubenswrapper[4806]: W1127 10:51:46.040546 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc358c0e7_2566_4f18_8a4a_2af284af91f6.slice/crio-0e8b22695b14f2d44c4684fb2b88837325e8844a10f570aec68cfcebc7581f5a WatchSource:0}: Error finding container 0e8b22695b14f2d44c4684fb2b88837325e8844a10f570aec68cfcebc7581f5a: Status 404 returned error can't find the container with id 0e8b22695b14f2d44c4684fb2b88837325e8844a10f570aec68cfcebc7581f5a Nov 27 10:51:47 crc kubenswrapper[4806]: I1127 10:51:47.039329 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" event={"ID":"c358c0e7-2566-4f18-8a4a-2af284af91f6","Type":"ContainerStarted","Data":"46433b167be5d0555b3b0b9c251a5c39782705ed33763a9ab51ec57d229fedaf"} Nov 27 10:51:47 crc kubenswrapper[4806]: I1127 10:51:47.039637 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" event={"ID":"c358c0e7-2566-4f18-8a4a-2af284af91f6","Type":"ContainerStarted","Data":"0e8b22695b14f2d44c4684fb2b88837325e8844a10f570aec68cfcebc7581f5a"} Nov 27 10:51:47 crc kubenswrapper[4806]: I1127 10:51:47.061913 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" podStartSLOduration=1.365334807 podStartE2EDuration="2.061893068s" podCreationTimestamp="2025-11-27 10:51:45 +0000 UTC" firstStartedPulling="2025-11-27 10:51:46.043872746 +0000 UTC m=+1810.630463510" lastFinishedPulling="2025-11-27 10:51:46.740431007 +0000 UTC m=+1811.327021771" observedRunningTime="2025-11-27 10:51:47.060530501 +0000 UTC m=+1811.647121265" watchObservedRunningTime="2025-11-27 10:51:47.061893068 +0000 UTC m=+1811.648483832" Nov 27 10:51:54 crc kubenswrapper[4806]: I1127 10:51:54.153469 4806 scope.go:117] "RemoveContainer" containerID="c3fe6f748b6629c4e9ba08c580894b58d63e5c15f1fc6e5887449b20e403ce38" Nov 27 10:51:57 crc kubenswrapper[4806]: I1127 10:51:57.149306 4806 generic.go:334] "Generic (PLEG): container finished" podID="c358c0e7-2566-4f18-8a4a-2af284af91f6" containerID="46433b167be5d0555b3b0b9c251a5c39782705ed33763a9ab51ec57d229fedaf" exitCode=0 Nov 27 10:51:57 crc kubenswrapper[4806]: I1127 10:51:57.150115 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" event={"ID":"c358c0e7-2566-4f18-8a4a-2af284af91f6","Type":"ContainerDied","Data":"46433b167be5d0555b3b0b9c251a5c39782705ed33763a9ab51ec57d229fedaf"} Nov 27 10:51:58 crc kubenswrapper[4806]: I1127 10:51:58.733774 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" Nov 27 10:51:58 crc kubenswrapper[4806]: I1127 10:51:58.808137 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxr2q\" (UniqueName: \"kubernetes.io/projected/c358c0e7-2566-4f18-8a4a-2af284af91f6-kube-api-access-mxr2q\") pod \"c358c0e7-2566-4f18-8a4a-2af284af91f6\" (UID: \"c358c0e7-2566-4f18-8a4a-2af284af91f6\") " Nov 27 10:51:58 crc kubenswrapper[4806]: I1127 10:51:58.808331 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c358c0e7-2566-4f18-8a4a-2af284af91f6-inventory\") pod \"c358c0e7-2566-4f18-8a4a-2af284af91f6\" (UID: \"c358c0e7-2566-4f18-8a4a-2af284af91f6\") " Nov 27 10:51:58 crc kubenswrapper[4806]: I1127 10:51:58.808417 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c358c0e7-2566-4f18-8a4a-2af284af91f6-ssh-key\") pod \"c358c0e7-2566-4f18-8a4a-2af284af91f6\" (UID: \"c358c0e7-2566-4f18-8a4a-2af284af91f6\") " Nov 27 10:51:58 crc kubenswrapper[4806]: I1127 10:51:58.818330 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c358c0e7-2566-4f18-8a4a-2af284af91f6-kube-api-access-mxr2q" (OuterVolumeSpecName: "kube-api-access-mxr2q") pod "c358c0e7-2566-4f18-8a4a-2af284af91f6" (UID: "c358c0e7-2566-4f18-8a4a-2af284af91f6"). InnerVolumeSpecName "kube-api-access-mxr2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:51:58 crc kubenswrapper[4806]: I1127 10:51:58.841646 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c358c0e7-2566-4f18-8a4a-2af284af91f6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c358c0e7-2566-4f18-8a4a-2af284af91f6" (UID: "c358c0e7-2566-4f18-8a4a-2af284af91f6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:51:58 crc kubenswrapper[4806]: I1127 10:51:58.846457 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c358c0e7-2566-4f18-8a4a-2af284af91f6-inventory" (OuterVolumeSpecName: "inventory") pod "c358c0e7-2566-4f18-8a4a-2af284af91f6" (UID: "c358c0e7-2566-4f18-8a4a-2af284af91f6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:51:58 crc kubenswrapper[4806]: I1127 10:51:58.910782 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c358c0e7-2566-4f18-8a4a-2af284af91f6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 10:51:58 crc kubenswrapper[4806]: I1127 10:51:58.911069 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxr2q\" (UniqueName: \"kubernetes.io/projected/c358c0e7-2566-4f18-8a4a-2af284af91f6-kube-api-access-mxr2q\") on node \"crc\" DevicePath \"\"" Nov 27 10:51:58 crc kubenswrapper[4806]: I1127 10:51:58.911308 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c358c0e7-2566-4f18-8a4a-2af284af91f6-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.186993 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" event={"ID":"c358c0e7-2566-4f18-8a4a-2af284af91f6","Type":"ContainerDied","Data":"0e8b22695b14f2d44c4684fb2b88837325e8844a10f570aec68cfcebc7581f5a"} Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.187048 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qvprb" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.187056 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e8b22695b14f2d44c4684fb2b88837325e8844a10f570aec68cfcebc7581f5a" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.269675 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td"] Nov 27 10:51:59 crc kubenswrapper[4806]: E1127 10:51:59.270606 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c358c0e7-2566-4f18-8a4a-2af284af91f6" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.270695 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c358c0e7-2566-4f18-8a4a-2af284af91f6" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.271068 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="c358c0e7-2566-4f18-8a4a-2af284af91f6" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.272026 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.279056 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.279064 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.279769 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.279795 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-j9ghp" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.280780 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td"] Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.319694 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp2tz\" (UniqueName: \"kubernetes.io/projected/078a0300-a31f-4b03-ba13-66692b945cb2-kube-api-access-tp2tz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z75td\" (UID: \"078a0300-a31f-4b03-ba13-66692b945cb2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.320180 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/078a0300-a31f-4b03-ba13-66692b945cb2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z75td\" (UID: \"078a0300-a31f-4b03-ba13-66692b945cb2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.320428 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/078a0300-a31f-4b03-ba13-66692b945cb2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z75td\" (UID: \"078a0300-a31f-4b03-ba13-66692b945cb2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.422685 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp2tz\" (UniqueName: \"kubernetes.io/projected/078a0300-a31f-4b03-ba13-66692b945cb2-kube-api-access-tp2tz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z75td\" (UID: \"078a0300-a31f-4b03-ba13-66692b945cb2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.423286 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/078a0300-a31f-4b03-ba13-66692b945cb2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z75td\" (UID: \"078a0300-a31f-4b03-ba13-66692b945cb2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.423410 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/078a0300-a31f-4b03-ba13-66692b945cb2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z75td\" (UID: \"078a0300-a31f-4b03-ba13-66692b945cb2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.428737 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/078a0300-a31f-4b03-ba13-66692b945cb2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z75td\" (UID: \"078a0300-a31f-4b03-ba13-66692b945cb2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.431203 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/078a0300-a31f-4b03-ba13-66692b945cb2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z75td\" (UID: \"078a0300-a31f-4b03-ba13-66692b945cb2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.445246 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp2tz\" (UniqueName: \"kubernetes.io/projected/078a0300-a31f-4b03-ba13-66692b945cb2-kube-api-access-tp2tz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-z75td\" (UID: \"078a0300-a31f-4b03-ba13-66692b945cb2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" Nov 27 10:51:59 crc kubenswrapper[4806]: I1127 10:51:59.595478 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" Nov 27 10:52:00 crc kubenswrapper[4806]: I1127 10:52:00.151864 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td"] Nov 27 10:52:00 crc kubenswrapper[4806]: I1127 10:52:00.197265 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" event={"ID":"078a0300-a31f-4b03-ba13-66692b945cb2","Type":"ContainerStarted","Data":"bbc010a526298167b8f41705cde78f5f3661e782c9cf625d5289ac50f9bcb4c2"} Nov 27 10:52:02 crc kubenswrapper[4806]: I1127 10:52:02.223026 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" event={"ID":"078a0300-a31f-4b03-ba13-66692b945cb2","Type":"ContainerStarted","Data":"268ffbb37429c6fb7816b8b750566ed79841fc71336bcdfd5fd1877da7d36360"} Nov 27 10:52:02 crc kubenswrapper[4806]: I1127 10:52:02.246006 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" podStartSLOduration=2.183570493 podStartE2EDuration="3.245975464s" podCreationTimestamp="2025-11-27 10:51:59 +0000 UTC" firstStartedPulling="2025-11-27 10:52:00.159856638 +0000 UTC m=+1824.746447402" lastFinishedPulling="2025-11-27 10:52:01.222261599 +0000 UTC m=+1825.808852373" observedRunningTime="2025-11-27 10:52:02.239075918 +0000 UTC m=+1826.825666682" watchObservedRunningTime="2025-11-27 10:52:02.245975464 +0000 UTC m=+1826.832566228" Nov 27 10:52:14 crc kubenswrapper[4806]: I1127 10:52:14.333908 4806 generic.go:334] "Generic (PLEG): container finished" podID="078a0300-a31f-4b03-ba13-66692b945cb2" containerID="268ffbb37429c6fb7816b8b750566ed79841fc71336bcdfd5fd1877da7d36360" exitCode=0 Nov 27 10:52:14 crc kubenswrapper[4806]: I1127 10:52:14.334701 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" event={"ID":"078a0300-a31f-4b03-ba13-66692b945cb2","Type":"ContainerDied","Data":"268ffbb37429c6fb7816b8b750566ed79841fc71336bcdfd5fd1877da7d36360"} Nov 27 10:52:15 crc kubenswrapper[4806]: I1127 10:52:15.770632 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" Nov 27 10:52:15 crc kubenswrapper[4806]: I1127 10:52:15.807152 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/078a0300-a31f-4b03-ba13-66692b945cb2-inventory\") pod \"078a0300-a31f-4b03-ba13-66692b945cb2\" (UID: \"078a0300-a31f-4b03-ba13-66692b945cb2\") " Nov 27 10:52:15 crc kubenswrapper[4806]: I1127 10:52:15.807468 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp2tz\" (UniqueName: \"kubernetes.io/projected/078a0300-a31f-4b03-ba13-66692b945cb2-kube-api-access-tp2tz\") pod \"078a0300-a31f-4b03-ba13-66692b945cb2\" (UID: \"078a0300-a31f-4b03-ba13-66692b945cb2\") " Nov 27 10:52:15 crc kubenswrapper[4806]: I1127 10:52:15.807503 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/078a0300-a31f-4b03-ba13-66692b945cb2-ssh-key\") pod \"078a0300-a31f-4b03-ba13-66692b945cb2\" (UID: \"078a0300-a31f-4b03-ba13-66692b945cb2\") " Nov 27 10:52:15 crc kubenswrapper[4806]: I1127 10:52:15.816827 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/078a0300-a31f-4b03-ba13-66692b945cb2-kube-api-access-tp2tz" (OuterVolumeSpecName: "kube-api-access-tp2tz") pod "078a0300-a31f-4b03-ba13-66692b945cb2" (UID: "078a0300-a31f-4b03-ba13-66692b945cb2"). InnerVolumeSpecName "kube-api-access-tp2tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:52:15 crc kubenswrapper[4806]: I1127 10:52:15.839555 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/078a0300-a31f-4b03-ba13-66692b945cb2-inventory" (OuterVolumeSpecName: "inventory") pod "078a0300-a31f-4b03-ba13-66692b945cb2" (UID: "078a0300-a31f-4b03-ba13-66692b945cb2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:52:15 crc kubenswrapper[4806]: I1127 10:52:15.844366 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/078a0300-a31f-4b03-ba13-66692b945cb2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "078a0300-a31f-4b03-ba13-66692b945cb2" (UID: "078a0300-a31f-4b03-ba13-66692b945cb2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 10:52:15 crc kubenswrapper[4806]: I1127 10:52:15.909784 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/078a0300-a31f-4b03-ba13-66692b945cb2-inventory\") on node \"crc\" DevicePath \"\"" Nov 27 10:52:15 crc kubenswrapper[4806]: I1127 10:52:15.909828 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp2tz\" (UniqueName: \"kubernetes.io/projected/078a0300-a31f-4b03-ba13-66692b945cb2-kube-api-access-tp2tz\") on node \"crc\" DevicePath \"\"" Nov 27 10:52:15 crc kubenswrapper[4806]: I1127 10:52:15.909844 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/078a0300-a31f-4b03-ba13-66692b945cb2-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 27 10:52:16 crc kubenswrapper[4806]: I1127 10:52:16.356209 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" event={"ID":"078a0300-a31f-4b03-ba13-66692b945cb2","Type":"ContainerDied","Data":"bbc010a526298167b8f41705cde78f5f3661e782c9cf625d5289ac50f9bcb4c2"} Nov 27 10:52:16 crc kubenswrapper[4806]: I1127 10:52:16.356285 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbc010a526298167b8f41705cde78f5f3661e782c9cf625d5289ac50f9bcb4c2" Nov 27 10:52:16 crc kubenswrapper[4806]: I1127 10:52:16.356395 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-z75td" Nov 27 10:53:44 crc kubenswrapper[4806]: I1127 10:53:44.450503 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:53:44 crc kubenswrapper[4806]: I1127 10:53:44.451127 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:54:14 crc kubenswrapper[4806]: I1127 10:54:14.450456 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:54:14 crc kubenswrapper[4806]: I1127 10:54:14.451295 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:54:44 crc kubenswrapper[4806]: I1127 10:54:44.450916 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:54:44 crc kubenswrapper[4806]: I1127 10:54:44.451748 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:54:44 crc kubenswrapper[4806]: I1127 10:54:44.451819 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:54:44 crc kubenswrapper[4806]: I1127 10:54:44.453152 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4499112e23b5522078d051c09243e588431a647460adb8b3e111de397927146b"} pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 10:54:44 crc kubenswrapper[4806]: I1127 10:54:44.453296 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" containerID="cri-o://4499112e23b5522078d051c09243e588431a647460adb8b3e111de397927146b" gracePeriod=600 Nov 27 10:54:45 crc kubenswrapper[4806]: I1127 10:54:45.877977 4806 generic.go:334] "Generic (PLEG): container finished" podID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerID="4499112e23b5522078d051c09243e588431a647460adb8b3e111de397927146b" exitCode=0 Nov 27 10:54:45 crc kubenswrapper[4806]: I1127 10:54:45.878026 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerDied","Data":"4499112e23b5522078d051c09243e588431a647460adb8b3e111de397927146b"} Nov 27 10:54:45 crc kubenswrapper[4806]: I1127 10:54:45.878813 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc"} Nov 27 10:54:45 crc kubenswrapper[4806]: I1127 10:54:45.878839 4806 scope.go:117] "RemoveContainer" containerID="659174b7253de14c2bd82f3b82d5293f867c48568c7951dbee5a335e879fd4a2" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.151736 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x44hk"] Nov 27 10:55:56 crc kubenswrapper[4806]: E1127 10:55:56.152814 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078a0300-a31f-4b03-ba13-66692b945cb2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.152835 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="078a0300-a31f-4b03-ba13-66692b945cb2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.153121 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="078a0300-a31f-4b03-ba13-66692b945cb2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.155409 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.167880 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x44hk"] Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.259168 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqkz8\" (UniqueName: \"kubernetes.io/projected/6a91920c-9638-45d7-8cdd-57b48d2acb0e-kube-api-access-sqkz8\") pod \"redhat-marketplace-x44hk\" (UID: \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\") " pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.259295 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a91920c-9638-45d7-8cdd-57b48d2acb0e-catalog-content\") pod \"redhat-marketplace-x44hk\" (UID: \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\") " pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.259344 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a91920c-9638-45d7-8cdd-57b48d2acb0e-utilities\") pod \"redhat-marketplace-x44hk\" (UID: \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\") " pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.361478 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqkz8\" (UniqueName: \"kubernetes.io/projected/6a91920c-9638-45d7-8cdd-57b48d2acb0e-kube-api-access-sqkz8\") pod \"redhat-marketplace-x44hk\" (UID: \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\") " pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.361570 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a91920c-9638-45d7-8cdd-57b48d2acb0e-catalog-content\") pod \"redhat-marketplace-x44hk\" (UID: \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\") " pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.361612 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a91920c-9638-45d7-8cdd-57b48d2acb0e-utilities\") pod \"redhat-marketplace-x44hk\" (UID: \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\") " pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.362135 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a91920c-9638-45d7-8cdd-57b48d2acb0e-utilities\") pod \"redhat-marketplace-x44hk\" (UID: \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\") " pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.362215 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a91920c-9638-45d7-8cdd-57b48d2acb0e-catalog-content\") pod \"redhat-marketplace-x44hk\" (UID: \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\") " pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.395008 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqkz8\" (UniqueName: \"kubernetes.io/projected/6a91920c-9638-45d7-8cdd-57b48d2acb0e-kube-api-access-sqkz8\") pod \"redhat-marketplace-x44hk\" (UID: \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\") " pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:55:56 crc kubenswrapper[4806]: I1127 10:55:56.491487 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:55:57 crc kubenswrapper[4806]: I1127 10:55:57.074485 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x44hk"] Nov 27 10:55:57 crc kubenswrapper[4806]: I1127 10:55:57.659978 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44hk" event={"ID":"6a91920c-9638-45d7-8cdd-57b48d2acb0e","Type":"ContainerStarted","Data":"26a3bd515b9ebc45f8a3ae4c94dc040d0cb2308f60cbf4dbccfe0933a32425e8"} Nov 27 10:55:57 crc kubenswrapper[4806]: I1127 10:55:57.660041 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44hk" event={"ID":"6a91920c-9638-45d7-8cdd-57b48d2acb0e","Type":"ContainerStarted","Data":"a5f43d9b8b7f27ef1796d63a55cfa99bf40fd9f88abfc12fd646c078f13e11cf"} Nov 27 10:55:58 crc kubenswrapper[4806]: I1127 10:55:58.671331 4806 generic.go:334] "Generic (PLEG): container finished" podID="6a91920c-9638-45d7-8cdd-57b48d2acb0e" containerID="26a3bd515b9ebc45f8a3ae4c94dc040d0cb2308f60cbf4dbccfe0933a32425e8" exitCode=0 Nov 27 10:55:58 crc kubenswrapper[4806]: I1127 10:55:58.671441 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44hk" event={"ID":"6a91920c-9638-45d7-8cdd-57b48d2acb0e","Type":"ContainerDied","Data":"26a3bd515b9ebc45f8a3ae4c94dc040d0cb2308f60cbf4dbccfe0933a32425e8"} Nov 27 10:55:58 crc kubenswrapper[4806]: I1127 10:55:58.674062 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 10:55:59 crc kubenswrapper[4806]: I1127 10:55:59.684745 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44hk" event={"ID":"6a91920c-9638-45d7-8cdd-57b48d2acb0e","Type":"ContainerStarted","Data":"0c3a4f011d0385849217ba5b925ba715944fa4f4ebb737636d27ffd2741aa95d"} Nov 27 10:56:00 crc kubenswrapper[4806]: I1127 10:56:00.698522 4806 generic.go:334] "Generic (PLEG): container finished" podID="6a91920c-9638-45d7-8cdd-57b48d2acb0e" containerID="0c3a4f011d0385849217ba5b925ba715944fa4f4ebb737636d27ffd2741aa95d" exitCode=0 Nov 27 10:56:00 crc kubenswrapper[4806]: I1127 10:56:00.698639 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44hk" event={"ID":"6a91920c-9638-45d7-8cdd-57b48d2acb0e","Type":"ContainerDied","Data":"0c3a4f011d0385849217ba5b925ba715944fa4f4ebb737636d27ffd2741aa95d"} Nov 27 10:56:02 crc kubenswrapper[4806]: I1127 10:56:02.736125 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44hk" event={"ID":"6a91920c-9638-45d7-8cdd-57b48d2acb0e","Type":"ContainerStarted","Data":"8997e1e8cdf0c1af00ee3eac98ab1e12962f58d6d1ea15f04d6f0904d19a572e"} Nov 27 10:56:02 crc kubenswrapper[4806]: I1127 10:56:02.766788 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x44hk" podStartSLOduration=3.713651558 podStartE2EDuration="6.766768066s" podCreationTimestamp="2025-11-27 10:55:56 +0000 UTC" firstStartedPulling="2025-11-27 10:55:58.673740798 +0000 UTC m=+2063.260331562" lastFinishedPulling="2025-11-27 10:56:01.726857306 +0000 UTC m=+2066.313448070" observedRunningTime="2025-11-27 10:56:02.76104066 +0000 UTC m=+2067.347631434" watchObservedRunningTime="2025-11-27 10:56:02.766768066 +0000 UTC m=+2067.353358830" Nov 27 10:56:06 crc kubenswrapper[4806]: I1127 10:56:06.491789 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:56:06 crc kubenswrapper[4806]: I1127 10:56:06.492496 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:56:06 crc kubenswrapper[4806]: I1127 10:56:06.569576 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:56:06 crc kubenswrapper[4806]: I1127 10:56:06.820419 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:56:06 crc kubenswrapper[4806]: I1127 10:56:06.886276 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x44hk"] Nov 27 10:56:08 crc kubenswrapper[4806]: I1127 10:56:08.794524 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x44hk" podUID="6a91920c-9638-45d7-8cdd-57b48d2acb0e" containerName="registry-server" containerID="cri-o://8997e1e8cdf0c1af00ee3eac98ab1e12962f58d6d1ea15f04d6f0904d19a572e" gracePeriod=2 Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.299912 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.376826 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqkz8\" (UniqueName: \"kubernetes.io/projected/6a91920c-9638-45d7-8cdd-57b48d2acb0e-kube-api-access-sqkz8\") pod \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\" (UID: \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\") " Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.377048 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a91920c-9638-45d7-8cdd-57b48d2acb0e-catalog-content\") pod \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\" (UID: \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\") " Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.377081 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a91920c-9638-45d7-8cdd-57b48d2acb0e-utilities\") pod \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\" (UID: \"6a91920c-9638-45d7-8cdd-57b48d2acb0e\") " Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.378466 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a91920c-9638-45d7-8cdd-57b48d2acb0e-utilities" (OuterVolumeSpecName: "utilities") pod "6a91920c-9638-45d7-8cdd-57b48d2acb0e" (UID: "6a91920c-9638-45d7-8cdd-57b48d2acb0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.383148 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a91920c-9638-45d7-8cdd-57b48d2acb0e-kube-api-access-sqkz8" (OuterVolumeSpecName: "kube-api-access-sqkz8") pod "6a91920c-9638-45d7-8cdd-57b48d2acb0e" (UID: "6a91920c-9638-45d7-8cdd-57b48d2acb0e"). InnerVolumeSpecName "kube-api-access-sqkz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.398422 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a91920c-9638-45d7-8cdd-57b48d2acb0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a91920c-9638-45d7-8cdd-57b48d2acb0e" (UID: "6a91920c-9638-45d7-8cdd-57b48d2acb0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.480357 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a91920c-9638-45d7-8cdd-57b48d2acb0e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.480406 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a91920c-9638-45d7-8cdd-57b48d2acb0e-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.480420 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqkz8\" (UniqueName: \"kubernetes.io/projected/6a91920c-9638-45d7-8cdd-57b48d2acb0e-kube-api-access-sqkz8\") on node \"crc\" DevicePath \"\"" Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.818877 4806 generic.go:334] "Generic (PLEG): container finished" podID="6a91920c-9638-45d7-8cdd-57b48d2acb0e" containerID="8997e1e8cdf0c1af00ee3eac98ab1e12962f58d6d1ea15f04d6f0904d19a572e" exitCode=0 Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.818944 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x44hk" Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.818958 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44hk" event={"ID":"6a91920c-9638-45d7-8cdd-57b48d2acb0e","Type":"ContainerDied","Data":"8997e1e8cdf0c1af00ee3eac98ab1e12962f58d6d1ea15f04d6f0904d19a572e"} Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.819001 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44hk" event={"ID":"6a91920c-9638-45d7-8cdd-57b48d2acb0e","Type":"ContainerDied","Data":"a5f43d9b8b7f27ef1796d63a55cfa99bf40fd9f88abfc12fd646c078f13e11cf"} Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.819033 4806 scope.go:117] "RemoveContainer" containerID="8997e1e8cdf0c1af00ee3eac98ab1e12962f58d6d1ea15f04d6f0904d19a572e" Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.870718 4806 scope.go:117] "RemoveContainer" containerID="0c3a4f011d0385849217ba5b925ba715944fa4f4ebb737636d27ffd2741aa95d" Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.924373 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x44hk"] Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.955821 4806 scope.go:117] "RemoveContainer" containerID="26a3bd515b9ebc45f8a3ae4c94dc040d0cb2308f60cbf4dbccfe0933a32425e8" Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.956918 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x44hk"] Nov 27 10:56:09 crc kubenswrapper[4806]: I1127 10:56:09.994763 4806 scope.go:117] "RemoveContainer" containerID="8997e1e8cdf0c1af00ee3eac98ab1e12962f58d6d1ea15f04d6f0904d19a572e" Nov 27 10:56:10 crc kubenswrapper[4806]: E1127 10:56:10.006894 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8997e1e8cdf0c1af00ee3eac98ab1e12962f58d6d1ea15f04d6f0904d19a572e\": container with ID starting with 8997e1e8cdf0c1af00ee3eac98ab1e12962f58d6d1ea15f04d6f0904d19a572e not found: ID does not exist" containerID="8997e1e8cdf0c1af00ee3eac98ab1e12962f58d6d1ea15f04d6f0904d19a572e" Nov 27 10:56:10 crc kubenswrapper[4806]: I1127 10:56:10.006958 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8997e1e8cdf0c1af00ee3eac98ab1e12962f58d6d1ea15f04d6f0904d19a572e"} err="failed to get container status \"8997e1e8cdf0c1af00ee3eac98ab1e12962f58d6d1ea15f04d6f0904d19a572e\": rpc error: code = NotFound desc = could not find container \"8997e1e8cdf0c1af00ee3eac98ab1e12962f58d6d1ea15f04d6f0904d19a572e\": container with ID starting with 8997e1e8cdf0c1af00ee3eac98ab1e12962f58d6d1ea15f04d6f0904d19a572e not found: ID does not exist" Nov 27 10:56:10 crc kubenswrapper[4806]: I1127 10:56:10.007002 4806 scope.go:117] "RemoveContainer" containerID="0c3a4f011d0385849217ba5b925ba715944fa4f4ebb737636d27ffd2741aa95d" Nov 27 10:56:10 crc kubenswrapper[4806]: E1127 10:56:10.012931 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c3a4f011d0385849217ba5b925ba715944fa4f4ebb737636d27ffd2741aa95d\": container with ID starting with 0c3a4f011d0385849217ba5b925ba715944fa4f4ebb737636d27ffd2741aa95d not found: ID does not exist" containerID="0c3a4f011d0385849217ba5b925ba715944fa4f4ebb737636d27ffd2741aa95d" Nov 27 10:56:10 crc kubenswrapper[4806]: I1127 10:56:10.013397 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c3a4f011d0385849217ba5b925ba715944fa4f4ebb737636d27ffd2741aa95d"} err="failed to get container status \"0c3a4f011d0385849217ba5b925ba715944fa4f4ebb737636d27ffd2741aa95d\": rpc error: code = NotFound desc = could not find container \"0c3a4f011d0385849217ba5b925ba715944fa4f4ebb737636d27ffd2741aa95d\": container with ID starting with 0c3a4f011d0385849217ba5b925ba715944fa4f4ebb737636d27ffd2741aa95d not found: ID does not exist" Nov 27 10:56:10 crc kubenswrapper[4806]: I1127 10:56:10.013435 4806 scope.go:117] "RemoveContainer" containerID="26a3bd515b9ebc45f8a3ae4c94dc040d0cb2308f60cbf4dbccfe0933a32425e8" Nov 27 10:56:10 crc kubenswrapper[4806]: E1127 10:56:10.018608 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26a3bd515b9ebc45f8a3ae4c94dc040d0cb2308f60cbf4dbccfe0933a32425e8\": container with ID starting with 26a3bd515b9ebc45f8a3ae4c94dc040d0cb2308f60cbf4dbccfe0933a32425e8 not found: ID does not exist" containerID="26a3bd515b9ebc45f8a3ae4c94dc040d0cb2308f60cbf4dbccfe0933a32425e8" Nov 27 10:56:10 crc kubenswrapper[4806]: I1127 10:56:10.018677 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26a3bd515b9ebc45f8a3ae4c94dc040d0cb2308f60cbf4dbccfe0933a32425e8"} err="failed to get container status \"26a3bd515b9ebc45f8a3ae4c94dc040d0cb2308f60cbf4dbccfe0933a32425e8\": rpc error: code = NotFound desc = could not find container \"26a3bd515b9ebc45f8a3ae4c94dc040d0cb2308f60cbf4dbccfe0933a32425e8\": container with ID starting with 26a3bd515b9ebc45f8a3ae4c94dc040d0cb2308f60cbf4dbccfe0933a32425e8 not found: ID does not exist" Nov 27 10:56:10 crc kubenswrapper[4806]: I1127 10:56:10.128124 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a91920c-9638-45d7-8cdd-57b48d2acb0e" path="/var/lib/kubelet/pods/6a91920c-9638-45d7-8cdd-57b48d2acb0e/volumes" Nov 27 10:57:14 crc kubenswrapper[4806]: I1127 10:57:14.450968 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:57:14 crc kubenswrapper[4806]: I1127 10:57:14.453560 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:57:44 crc kubenswrapper[4806]: I1127 10:57:44.450787 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:57:44 crc kubenswrapper[4806]: I1127 10:57:44.451759 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.450275 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.451302 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.451379 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.452206 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc"} pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.452288 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" containerID="cri-o://90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" gracePeriod=600 Nov 27 10:58:14 crc kubenswrapper[4806]: E1127 10:58:14.603289 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.605817 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xsq8v"] Nov 27 10:58:14 crc kubenswrapper[4806]: E1127 10:58:14.606323 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a91920c-9638-45d7-8cdd-57b48d2acb0e" containerName="extract-utilities" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.606343 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a91920c-9638-45d7-8cdd-57b48d2acb0e" containerName="extract-utilities" Nov 27 10:58:14 crc kubenswrapper[4806]: E1127 10:58:14.606352 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a91920c-9638-45d7-8cdd-57b48d2acb0e" containerName="extract-content" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.606360 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a91920c-9638-45d7-8cdd-57b48d2acb0e" containerName="extract-content" Nov 27 10:58:14 crc kubenswrapper[4806]: E1127 10:58:14.606404 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a91920c-9638-45d7-8cdd-57b48d2acb0e" containerName="registry-server" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.606412 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a91920c-9638-45d7-8cdd-57b48d2acb0e" containerName="registry-server" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.606600 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a91920c-9638-45d7-8cdd-57b48d2acb0e" containerName="registry-server" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.612806 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.629603 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xsq8v"] Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.756195 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/611aba0f-c497-4a3b-978a-187fc9f325b7-utilities\") pod \"certified-operators-xsq8v\" (UID: \"611aba0f-c497-4a3b-978a-187fc9f325b7\") " pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.756813 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/611aba0f-c497-4a3b-978a-187fc9f325b7-catalog-content\") pod \"certified-operators-xsq8v\" (UID: \"611aba0f-c497-4a3b-978a-187fc9f325b7\") " pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.757063 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9cvd\" (UniqueName: \"kubernetes.io/projected/611aba0f-c497-4a3b-978a-187fc9f325b7-kube-api-access-l9cvd\") pod \"certified-operators-xsq8v\" (UID: \"611aba0f-c497-4a3b-978a-187fc9f325b7\") " pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.859387 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/611aba0f-c497-4a3b-978a-187fc9f325b7-utilities\") pod \"certified-operators-xsq8v\" (UID: \"611aba0f-c497-4a3b-978a-187fc9f325b7\") " pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.859441 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/611aba0f-c497-4a3b-978a-187fc9f325b7-catalog-content\") pod \"certified-operators-xsq8v\" (UID: \"611aba0f-c497-4a3b-978a-187fc9f325b7\") " pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.859508 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9cvd\" (UniqueName: \"kubernetes.io/projected/611aba0f-c497-4a3b-978a-187fc9f325b7-kube-api-access-l9cvd\") pod \"certified-operators-xsq8v\" (UID: \"611aba0f-c497-4a3b-978a-187fc9f325b7\") " pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.860498 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/611aba0f-c497-4a3b-978a-187fc9f325b7-utilities\") pod \"certified-operators-xsq8v\" (UID: \"611aba0f-c497-4a3b-978a-187fc9f325b7\") " pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.860756 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/611aba0f-c497-4a3b-978a-187fc9f325b7-catalog-content\") pod \"certified-operators-xsq8v\" (UID: \"611aba0f-c497-4a3b-978a-187fc9f325b7\") " pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.887346 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9cvd\" (UniqueName: \"kubernetes.io/projected/611aba0f-c497-4a3b-978a-187fc9f325b7-kube-api-access-l9cvd\") pod \"certified-operators-xsq8v\" (UID: \"611aba0f-c497-4a3b-978a-187fc9f325b7\") " pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:14 crc kubenswrapper[4806]: I1127 10:58:14.943717 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:15 crc kubenswrapper[4806]: I1127 10:58:15.284698 4806 generic.go:334] "Generic (PLEG): container finished" podID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" exitCode=0 Nov 27 10:58:15 crc kubenswrapper[4806]: I1127 10:58:15.285111 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerDied","Data":"90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc"} Nov 27 10:58:15 crc kubenswrapper[4806]: I1127 10:58:15.285161 4806 scope.go:117] "RemoveContainer" containerID="4499112e23b5522078d051c09243e588431a647460adb8b3e111de397927146b" Nov 27 10:58:15 crc kubenswrapper[4806]: I1127 10:58:15.286031 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 10:58:15 crc kubenswrapper[4806]: E1127 10:58:15.286284 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:58:15 crc kubenswrapper[4806]: I1127 10:58:15.531654 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xsq8v"] Nov 27 10:58:16 crc kubenswrapper[4806]: I1127 10:58:16.305367 4806 generic.go:334] "Generic (PLEG): container finished" podID="611aba0f-c497-4a3b-978a-187fc9f325b7" containerID="23481c57be341d00e00be0b100bac95c0793bca3f3e54e819e68d223f6d5239b" exitCode=0 Nov 27 10:58:16 crc kubenswrapper[4806]: I1127 10:58:16.305484 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xsq8v" event={"ID":"611aba0f-c497-4a3b-978a-187fc9f325b7","Type":"ContainerDied","Data":"23481c57be341d00e00be0b100bac95c0793bca3f3e54e819e68d223f6d5239b"} Nov 27 10:58:16 crc kubenswrapper[4806]: I1127 10:58:16.306314 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xsq8v" event={"ID":"611aba0f-c497-4a3b-978a-187fc9f325b7","Type":"ContainerStarted","Data":"adb6c081e614bbd5f84e5b9c008b0d3872c8ec6ec55b3285a817a6c7843b2d18"} Nov 27 10:58:18 crc kubenswrapper[4806]: I1127 10:58:18.325964 4806 generic.go:334] "Generic (PLEG): container finished" podID="611aba0f-c497-4a3b-978a-187fc9f325b7" containerID="3ebc3fb742fbdabfd58531104500e9c32a701eb1f963d645198ae8ae7fa52dac" exitCode=0 Nov 27 10:58:18 crc kubenswrapper[4806]: I1127 10:58:18.326044 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xsq8v" event={"ID":"611aba0f-c497-4a3b-978a-187fc9f325b7","Type":"ContainerDied","Data":"3ebc3fb742fbdabfd58531104500e9c32a701eb1f963d645198ae8ae7fa52dac"} Nov 27 10:58:19 crc kubenswrapper[4806]: I1127 10:58:19.340283 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xsq8v" event={"ID":"611aba0f-c497-4a3b-978a-187fc9f325b7","Type":"ContainerStarted","Data":"67dd0ac86ad7eda4d564619129a7b61b47cd3cdd635c16ada0215ed0a8d59c26"} Nov 27 10:58:19 crc kubenswrapper[4806]: I1127 10:58:19.371084 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xsq8v" podStartSLOduration=2.93942431 podStartE2EDuration="5.371050807s" podCreationTimestamp="2025-11-27 10:58:14 +0000 UTC" firstStartedPulling="2025-11-27 10:58:16.308913274 +0000 UTC m=+2200.895504068" lastFinishedPulling="2025-11-27 10:58:18.740539791 +0000 UTC m=+2203.327130565" observedRunningTime="2025-11-27 10:58:19.360937072 +0000 UTC m=+2203.947527846" watchObservedRunningTime="2025-11-27 10:58:19.371050807 +0000 UTC m=+2203.957641571" Nov 27 10:58:24 crc kubenswrapper[4806]: I1127 10:58:24.944376 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:24 crc kubenswrapper[4806]: I1127 10:58:24.944747 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:24 crc kubenswrapper[4806]: I1127 10:58:24.998069 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:25 crc kubenswrapper[4806]: I1127 10:58:25.464085 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:25 crc kubenswrapper[4806]: I1127 10:58:25.529199 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xsq8v"] Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.430303 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xsq8v" podUID="611aba0f-c497-4a3b-978a-187fc9f325b7" containerName="registry-server" containerID="cri-o://67dd0ac86ad7eda4d564619129a7b61b47cd3cdd635c16ada0215ed0a8d59c26" gracePeriod=2 Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.679400 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dfhcb"] Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.683634 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.692773 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dfhcb"] Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.757853 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zvvs\" (UniqueName: \"kubernetes.io/projected/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-kube-api-access-7zvvs\") pod \"community-operators-dfhcb\" (UID: \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\") " pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.758090 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-utilities\") pod \"community-operators-dfhcb\" (UID: \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\") " pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.758299 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-catalog-content\") pod \"community-operators-dfhcb\" (UID: \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\") " pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.877979 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zvvs\" (UniqueName: \"kubernetes.io/projected/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-kube-api-access-7zvvs\") pod \"community-operators-dfhcb\" (UID: \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\") " pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.878094 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-utilities\") pod \"community-operators-dfhcb\" (UID: \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\") " pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.878215 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-catalog-content\") pod \"community-operators-dfhcb\" (UID: \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\") " pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.879283 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-catalog-content\") pod \"community-operators-dfhcb\" (UID: \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\") " pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.879484 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-utilities\") pod \"community-operators-dfhcb\" (UID: \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\") " pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.907194 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zvvs\" (UniqueName: \"kubernetes.io/projected/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-kube-api-access-7zvvs\") pod \"community-operators-dfhcb\" (UID: \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\") " pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:27 crc kubenswrapper[4806]: I1127 10:58:27.969528 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.029681 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.082068 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/611aba0f-c497-4a3b-978a-187fc9f325b7-utilities\") pod \"611aba0f-c497-4a3b-978a-187fc9f325b7\" (UID: \"611aba0f-c497-4a3b-978a-187fc9f325b7\") " Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.082147 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/611aba0f-c497-4a3b-978a-187fc9f325b7-catalog-content\") pod \"611aba0f-c497-4a3b-978a-187fc9f325b7\" (UID: \"611aba0f-c497-4a3b-978a-187fc9f325b7\") " Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.082383 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9cvd\" (UniqueName: \"kubernetes.io/projected/611aba0f-c497-4a3b-978a-187fc9f325b7-kube-api-access-l9cvd\") pod \"611aba0f-c497-4a3b-978a-187fc9f325b7\" (UID: \"611aba0f-c497-4a3b-978a-187fc9f325b7\") " Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.086509 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/611aba0f-c497-4a3b-978a-187fc9f325b7-utilities" (OuterVolumeSpecName: "utilities") pod "611aba0f-c497-4a3b-978a-187fc9f325b7" (UID: "611aba0f-c497-4a3b-978a-187fc9f325b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.093509 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/611aba0f-c497-4a3b-978a-187fc9f325b7-kube-api-access-l9cvd" (OuterVolumeSpecName: "kube-api-access-l9cvd") pod "611aba0f-c497-4a3b-978a-187fc9f325b7" (UID: "611aba0f-c497-4a3b-978a-187fc9f325b7"). InnerVolumeSpecName "kube-api-access-l9cvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.151521 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/611aba0f-c497-4a3b-978a-187fc9f325b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "611aba0f-c497-4a3b-978a-187fc9f325b7" (UID: "611aba0f-c497-4a3b-978a-187fc9f325b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.186533 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9cvd\" (UniqueName: \"kubernetes.io/projected/611aba0f-c497-4a3b-978a-187fc9f325b7-kube-api-access-l9cvd\") on node \"crc\" DevicePath \"\"" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.186578 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/611aba0f-c497-4a3b-978a-187fc9f325b7-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.186592 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/611aba0f-c497-4a3b-978a-187fc9f325b7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.477660 4806 generic.go:334] "Generic (PLEG): container finished" podID="611aba0f-c497-4a3b-978a-187fc9f325b7" containerID="67dd0ac86ad7eda4d564619129a7b61b47cd3cdd635c16ada0215ed0a8d59c26" exitCode=0 Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.477828 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xsq8v" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.477866 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xsq8v" event={"ID":"611aba0f-c497-4a3b-978a-187fc9f325b7","Type":"ContainerDied","Data":"67dd0ac86ad7eda4d564619129a7b61b47cd3cdd635c16ada0215ed0a8d59c26"} Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.483673 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xsq8v" event={"ID":"611aba0f-c497-4a3b-978a-187fc9f325b7","Type":"ContainerDied","Data":"adb6c081e614bbd5f84e5b9c008b0d3872c8ec6ec55b3285a817a6c7843b2d18"} Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.483811 4806 scope.go:117] "RemoveContainer" containerID="67dd0ac86ad7eda4d564619129a7b61b47cd3cdd635c16ada0215ed0a8d59c26" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.551034 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xsq8v"] Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.565586 4806 scope.go:117] "RemoveContainer" containerID="3ebc3fb742fbdabfd58531104500e9c32a701eb1f963d645198ae8ae7fa52dac" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.570625 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xsq8v"] Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.616551 4806 scope.go:117] "RemoveContainer" containerID="23481c57be341d00e00be0b100bac95c0793bca3f3e54e819e68d223f6d5239b" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.667165 4806 scope.go:117] "RemoveContainer" containerID="67dd0ac86ad7eda4d564619129a7b61b47cd3cdd635c16ada0215ed0a8d59c26" Nov 27 10:58:28 crc kubenswrapper[4806]: E1127 10:58:28.667969 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67dd0ac86ad7eda4d564619129a7b61b47cd3cdd635c16ada0215ed0a8d59c26\": container with ID starting with 67dd0ac86ad7eda4d564619129a7b61b47cd3cdd635c16ada0215ed0a8d59c26 not found: ID does not exist" containerID="67dd0ac86ad7eda4d564619129a7b61b47cd3cdd635c16ada0215ed0a8d59c26" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.668021 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67dd0ac86ad7eda4d564619129a7b61b47cd3cdd635c16ada0215ed0a8d59c26"} err="failed to get container status \"67dd0ac86ad7eda4d564619129a7b61b47cd3cdd635c16ada0215ed0a8d59c26\": rpc error: code = NotFound desc = could not find container \"67dd0ac86ad7eda4d564619129a7b61b47cd3cdd635c16ada0215ed0a8d59c26\": container with ID starting with 67dd0ac86ad7eda4d564619129a7b61b47cd3cdd635c16ada0215ed0a8d59c26 not found: ID does not exist" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.668052 4806 scope.go:117] "RemoveContainer" containerID="3ebc3fb742fbdabfd58531104500e9c32a701eb1f963d645198ae8ae7fa52dac" Nov 27 10:58:28 crc kubenswrapper[4806]: E1127 10:58:28.668790 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ebc3fb742fbdabfd58531104500e9c32a701eb1f963d645198ae8ae7fa52dac\": container with ID starting with 3ebc3fb742fbdabfd58531104500e9c32a701eb1f963d645198ae8ae7fa52dac not found: ID does not exist" containerID="3ebc3fb742fbdabfd58531104500e9c32a701eb1f963d645198ae8ae7fa52dac" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.668822 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ebc3fb742fbdabfd58531104500e9c32a701eb1f963d645198ae8ae7fa52dac"} err="failed to get container status \"3ebc3fb742fbdabfd58531104500e9c32a701eb1f963d645198ae8ae7fa52dac\": rpc error: code = NotFound desc = could not find container \"3ebc3fb742fbdabfd58531104500e9c32a701eb1f963d645198ae8ae7fa52dac\": container with ID starting with 3ebc3fb742fbdabfd58531104500e9c32a701eb1f963d645198ae8ae7fa52dac not found: ID does not exist" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.668841 4806 scope.go:117] "RemoveContainer" containerID="23481c57be341d00e00be0b100bac95c0793bca3f3e54e819e68d223f6d5239b" Nov 27 10:58:28 crc kubenswrapper[4806]: E1127 10:58:28.669505 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23481c57be341d00e00be0b100bac95c0793bca3f3e54e819e68d223f6d5239b\": container with ID starting with 23481c57be341d00e00be0b100bac95c0793bca3f3e54e819e68d223f6d5239b not found: ID does not exist" containerID="23481c57be341d00e00be0b100bac95c0793bca3f3e54e819e68d223f6d5239b" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.669545 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23481c57be341d00e00be0b100bac95c0793bca3f3e54e819e68d223f6d5239b"} err="failed to get container status \"23481c57be341d00e00be0b100bac95c0793bca3f3e54e819e68d223f6d5239b\": rpc error: code = NotFound desc = could not find container \"23481c57be341d00e00be0b100bac95c0793bca3f3e54e819e68d223f6d5239b\": container with ID starting with 23481c57be341d00e00be0b100bac95c0793bca3f3e54e819e68d223f6d5239b not found: ID does not exist" Nov 27 10:58:28 crc kubenswrapper[4806]: I1127 10:58:28.792748 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dfhcb"] Nov 27 10:58:29 crc kubenswrapper[4806]: I1127 10:58:29.488829 4806 generic.go:334] "Generic (PLEG): container finished" podID="ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" containerID="71ce11f3c0a3cc21f0290e826c5574d33438999283c7433d0cba9453a83a8bb2" exitCode=0 Nov 27 10:58:29 crc kubenswrapper[4806]: I1127 10:58:29.488917 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfhcb" event={"ID":"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6","Type":"ContainerDied","Data":"71ce11f3c0a3cc21f0290e826c5574d33438999283c7433d0cba9453a83a8bb2"} Nov 27 10:58:29 crc kubenswrapper[4806]: I1127 10:58:29.489112 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfhcb" event={"ID":"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6","Type":"ContainerStarted","Data":"12e0195f87dc7962bab026e9072d66b926ef1742b6ceb6a2619dba1aaf43fea4"} Nov 27 10:58:30 crc kubenswrapper[4806]: I1127 10:58:30.117506 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 10:58:30 crc kubenswrapper[4806]: E1127 10:58:30.118044 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:58:30 crc kubenswrapper[4806]: I1127 10:58:30.128970 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="611aba0f-c497-4a3b-978a-187fc9f325b7" path="/var/lib/kubelet/pods/611aba0f-c497-4a3b-978a-187fc9f325b7/volumes" Nov 27 10:58:30 crc kubenswrapper[4806]: I1127 10:58:30.506793 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfhcb" event={"ID":"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6","Type":"ContainerStarted","Data":"279e69cf523cc209f68ad74c132dfc6f1fc57fe72a3564ebbe15374dd14fb612"} Nov 27 10:58:32 crc kubenswrapper[4806]: I1127 10:58:32.530923 4806 generic.go:334] "Generic (PLEG): container finished" podID="ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" containerID="279e69cf523cc209f68ad74c132dfc6f1fc57fe72a3564ebbe15374dd14fb612" exitCode=0 Nov 27 10:58:32 crc kubenswrapper[4806]: I1127 10:58:32.531000 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfhcb" event={"ID":"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6","Type":"ContainerDied","Data":"279e69cf523cc209f68ad74c132dfc6f1fc57fe72a3564ebbe15374dd14fb612"} Nov 27 10:58:33 crc kubenswrapper[4806]: I1127 10:58:33.543195 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfhcb" event={"ID":"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6","Type":"ContainerStarted","Data":"c6cd68c59be0b3df5a0183e637b5918060c0b4f1e95fe55cc50648696bc12c64"} Nov 27 10:58:33 crc kubenswrapper[4806]: I1127 10:58:33.575822 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dfhcb" podStartSLOduration=3.147507721 podStartE2EDuration="6.575796017s" podCreationTimestamp="2025-11-27 10:58:27 +0000 UTC" firstStartedPulling="2025-11-27 10:58:29.491066931 +0000 UTC m=+2214.077657695" lastFinishedPulling="2025-11-27 10:58:32.919355217 +0000 UTC m=+2217.505945991" observedRunningTime="2025-11-27 10:58:33.571407045 +0000 UTC m=+2218.157997829" watchObservedRunningTime="2025-11-27 10:58:33.575796017 +0000 UTC m=+2218.162386781" Nov 27 10:58:38 crc kubenswrapper[4806]: I1127 10:58:38.029995 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:38 crc kubenswrapper[4806]: I1127 10:58:38.031878 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:38 crc kubenswrapper[4806]: I1127 10:58:38.080745 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:38 crc kubenswrapper[4806]: I1127 10:58:38.654175 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:38 crc kubenswrapper[4806]: I1127 10:58:38.728214 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dfhcb"] Nov 27 10:58:40 crc kubenswrapper[4806]: I1127 10:58:40.612284 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dfhcb" podUID="ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" containerName="registry-server" containerID="cri-o://c6cd68c59be0b3df5a0183e637b5918060c0b4f1e95fe55cc50648696bc12c64" gracePeriod=2 Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.120079 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.293678 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-utilities\") pod \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\" (UID: \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\") " Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.293752 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-catalog-content\") pod \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\" (UID: \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\") " Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.293921 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zvvs\" (UniqueName: \"kubernetes.io/projected/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-kube-api-access-7zvvs\") pod \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\" (UID: \"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6\") " Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.294715 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-utilities" (OuterVolumeSpecName: "utilities") pod "ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" (UID: "ee9839ca-59b3-4ef8-95fd-81bdff3cbab6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.303999 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-kube-api-access-7zvvs" (OuterVolumeSpecName: "kube-api-access-7zvvs") pod "ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" (UID: "ee9839ca-59b3-4ef8-95fd-81bdff3cbab6"). InnerVolumeSpecName "kube-api-access-7zvvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.358878 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" (UID: "ee9839ca-59b3-4ef8-95fd-81bdff3cbab6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.395184 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.395255 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.395272 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zvvs\" (UniqueName: \"kubernetes.io/projected/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6-kube-api-access-7zvvs\") on node \"crc\" DevicePath \"\"" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.630280 4806 generic.go:334] "Generic (PLEG): container finished" podID="ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" containerID="c6cd68c59be0b3df5a0183e637b5918060c0b4f1e95fe55cc50648696bc12c64" exitCode=0 Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.630344 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfhcb" event={"ID":"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6","Type":"ContainerDied","Data":"c6cd68c59be0b3df5a0183e637b5918060c0b4f1e95fe55cc50648696bc12c64"} Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.630388 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfhcb" event={"ID":"ee9839ca-59b3-4ef8-95fd-81bdff3cbab6","Type":"ContainerDied","Data":"12e0195f87dc7962bab026e9072d66b926ef1742b6ceb6a2619dba1aaf43fea4"} Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.630410 4806 scope.go:117] "RemoveContainer" containerID="c6cd68c59be0b3df5a0183e637b5918060c0b4f1e95fe55cc50648696bc12c64" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.630473 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dfhcb" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.660396 4806 scope.go:117] "RemoveContainer" containerID="279e69cf523cc209f68ad74c132dfc6f1fc57fe72a3564ebbe15374dd14fb612" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.695759 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dfhcb"] Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.705197 4806 scope.go:117] "RemoveContainer" containerID="71ce11f3c0a3cc21f0290e826c5574d33438999283c7433d0cba9453a83a8bb2" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.712281 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dfhcb"] Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.746348 4806 scope.go:117] "RemoveContainer" containerID="c6cd68c59be0b3df5a0183e637b5918060c0b4f1e95fe55cc50648696bc12c64" Nov 27 10:58:41 crc kubenswrapper[4806]: E1127 10:58:41.746929 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6cd68c59be0b3df5a0183e637b5918060c0b4f1e95fe55cc50648696bc12c64\": container with ID starting with c6cd68c59be0b3df5a0183e637b5918060c0b4f1e95fe55cc50648696bc12c64 not found: ID does not exist" containerID="c6cd68c59be0b3df5a0183e637b5918060c0b4f1e95fe55cc50648696bc12c64" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.746967 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6cd68c59be0b3df5a0183e637b5918060c0b4f1e95fe55cc50648696bc12c64"} err="failed to get container status \"c6cd68c59be0b3df5a0183e637b5918060c0b4f1e95fe55cc50648696bc12c64\": rpc error: code = NotFound desc = could not find container \"c6cd68c59be0b3df5a0183e637b5918060c0b4f1e95fe55cc50648696bc12c64\": container with ID starting with c6cd68c59be0b3df5a0183e637b5918060c0b4f1e95fe55cc50648696bc12c64 not found: ID does not exist" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.747000 4806 scope.go:117] "RemoveContainer" containerID="279e69cf523cc209f68ad74c132dfc6f1fc57fe72a3564ebbe15374dd14fb612" Nov 27 10:58:41 crc kubenswrapper[4806]: E1127 10:58:41.748008 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"279e69cf523cc209f68ad74c132dfc6f1fc57fe72a3564ebbe15374dd14fb612\": container with ID starting with 279e69cf523cc209f68ad74c132dfc6f1fc57fe72a3564ebbe15374dd14fb612 not found: ID does not exist" containerID="279e69cf523cc209f68ad74c132dfc6f1fc57fe72a3564ebbe15374dd14fb612" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.748033 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"279e69cf523cc209f68ad74c132dfc6f1fc57fe72a3564ebbe15374dd14fb612"} err="failed to get container status \"279e69cf523cc209f68ad74c132dfc6f1fc57fe72a3564ebbe15374dd14fb612\": rpc error: code = NotFound desc = could not find container \"279e69cf523cc209f68ad74c132dfc6f1fc57fe72a3564ebbe15374dd14fb612\": container with ID starting with 279e69cf523cc209f68ad74c132dfc6f1fc57fe72a3564ebbe15374dd14fb612 not found: ID does not exist" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.748045 4806 scope.go:117] "RemoveContainer" containerID="71ce11f3c0a3cc21f0290e826c5574d33438999283c7433d0cba9453a83a8bb2" Nov 27 10:58:41 crc kubenswrapper[4806]: E1127 10:58:41.748564 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71ce11f3c0a3cc21f0290e826c5574d33438999283c7433d0cba9453a83a8bb2\": container with ID starting with 71ce11f3c0a3cc21f0290e826c5574d33438999283c7433d0cba9453a83a8bb2 not found: ID does not exist" containerID="71ce11f3c0a3cc21f0290e826c5574d33438999283c7433d0cba9453a83a8bb2" Nov 27 10:58:41 crc kubenswrapper[4806]: I1127 10:58:41.748632 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71ce11f3c0a3cc21f0290e826c5574d33438999283c7433d0cba9453a83a8bb2"} err="failed to get container status \"71ce11f3c0a3cc21f0290e826c5574d33438999283c7433d0cba9453a83a8bb2\": rpc error: code = NotFound desc = could not find container \"71ce11f3c0a3cc21f0290e826c5574d33438999283c7433d0cba9453a83a8bb2\": container with ID starting with 71ce11f3c0a3cc21f0290e826c5574d33438999283c7433d0cba9453a83a8bb2 not found: ID does not exist" Nov 27 10:58:42 crc kubenswrapper[4806]: I1127 10:58:42.134177 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" path="/var/lib/kubelet/pods/ee9839ca-59b3-4ef8-95fd-81bdff3cbab6/volumes" Nov 27 10:58:44 crc kubenswrapper[4806]: I1127 10:58:44.118378 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 10:58:44 crc kubenswrapper[4806]: E1127 10:58:44.119403 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:58:58 crc kubenswrapper[4806]: I1127 10:58:58.117863 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 10:58:58 crc kubenswrapper[4806]: E1127 10:58:58.119160 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:59:11 crc kubenswrapper[4806]: I1127 10:59:11.117491 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 10:59:11 crc kubenswrapper[4806]: E1127 10:59:11.118414 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:59:26 crc kubenswrapper[4806]: I1127 10:59:26.125175 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 10:59:26 crc kubenswrapper[4806]: E1127 10:59:26.129683 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:59:39 crc kubenswrapper[4806]: I1127 10:59:39.117737 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 10:59:39 crc kubenswrapper[4806]: E1127 10:59:39.119008 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 10:59:51 crc kubenswrapper[4806]: I1127 10:59:51.118224 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 10:59:51 crc kubenswrapper[4806]: E1127 10:59:51.120894 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.189817 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx"] Nov 27 11:00:00 crc kubenswrapper[4806]: E1127 11:00:00.191647 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="611aba0f-c497-4a3b-978a-187fc9f325b7" containerName="extract-utilities" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.191666 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="611aba0f-c497-4a3b-978a-187fc9f325b7" containerName="extract-utilities" Nov 27 11:00:00 crc kubenswrapper[4806]: E1127 11:00:00.191679 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" containerName="extract-content" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.191687 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" containerName="extract-content" Nov 27 11:00:00 crc kubenswrapper[4806]: E1127 11:00:00.191707 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="611aba0f-c497-4a3b-978a-187fc9f325b7" containerName="extract-content" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.191714 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="611aba0f-c497-4a3b-978a-187fc9f325b7" containerName="extract-content" Nov 27 11:00:00 crc kubenswrapper[4806]: E1127 11:00:00.191726 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" containerName="registry-server" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.191733 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" containerName="registry-server" Nov 27 11:00:00 crc kubenswrapper[4806]: E1127 11:00:00.191754 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" containerName="extract-utilities" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.191761 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" containerName="extract-utilities" Nov 27 11:00:00 crc kubenswrapper[4806]: E1127 11:00:00.191779 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="611aba0f-c497-4a3b-978a-187fc9f325b7" containerName="registry-server" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.191785 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="611aba0f-c497-4a3b-978a-187fc9f325b7" containerName="registry-server" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.192009 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="611aba0f-c497-4a3b-978a-187fc9f325b7" containerName="registry-server" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.192217 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee9839ca-59b3-4ef8-95fd-81bdff3cbab6" containerName="registry-server" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.193141 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.200591 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.200634 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.202681 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx"] Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.280275 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkqq7\" (UniqueName: \"kubernetes.io/projected/9169c112-b10b-4938-bf73-c4f360d886ac-kube-api-access-xkqq7\") pod \"collect-profiles-29404020-mh5cx\" (UID: \"9169c112-b10b-4938-bf73-c4f360d886ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.280707 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9169c112-b10b-4938-bf73-c4f360d886ac-config-volume\") pod \"collect-profiles-29404020-mh5cx\" (UID: \"9169c112-b10b-4938-bf73-c4f360d886ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.280773 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9169c112-b10b-4938-bf73-c4f360d886ac-secret-volume\") pod \"collect-profiles-29404020-mh5cx\" (UID: \"9169c112-b10b-4938-bf73-c4f360d886ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.382206 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkqq7\" (UniqueName: \"kubernetes.io/projected/9169c112-b10b-4938-bf73-c4f360d886ac-kube-api-access-xkqq7\") pod \"collect-profiles-29404020-mh5cx\" (UID: \"9169c112-b10b-4938-bf73-c4f360d886ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.382292 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9169c112-b10b-4938-bf73-c4f360d886ac-config-volume\") pod \"collect-profiles-29404020-mh5cx\" (UID: \"9169c112-b10b-4938-bf73-c4f360d886ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.382337 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9169c112-b10b-4938-bf73-c4f360d886ac-secret-volume\") pod \"collect-profiles-29404020-mh5cx\" (UID: \"9169c112-b10b-4938-bf73-c4f360d886ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.384360 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9169c112-b10b-4938-bf73-c4f360d886ac-config-volume\") pod \"collect-profiles-29404020-mh5cx\" (UID: \"9169c112-b10b-4938-bf73-c4f360d886ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.389054 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9169c112-b10b-4938-bf73-c4f360d886ac-secret-volume\") pod \"collect-profiles-29404020-mh5cx\" (UID: \"9169c112-b10b-4938-bf73-c4f360d886ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.401100 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkqq7\" (UniqueName: \"kubernetes.io/projected/9169c112-b10b-4938-bf73-c4f360d886ac-kube-api-access-xkqq7\") pod \"collect-profiles-29404020-mh5cx\" (UID: \"9169c112-b10b-4938-bf73-c4f360d886ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" Nov 27 11:00:00 crc kubenswrapper[4806]: I1127 11:00:00.521655 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" Nov 27 11:00:01 crc kubenswrapper[4806]: I1127 11:00:01.056342 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx"] Nov 27 11:00:01 crc kubenswrapper[4806]: W1127 11:00:01.072423 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9169c112_b10b_4938_bf73_c4f360d886ac.slice/crio-73388ee0f606c4f1531f1bcf839c6569ec758c8dc19cc5921d7bcc6b8850f865 WatchSource:0}: Error finding container 73388ee0f606c4f1531f1bcf839c6569ec758c8dc19cc5921d7bcc6b8850f865: Status 404 returned error can't find the container with id 73388ee0f606c4f1531f1bcf839c6569ec758c8dc19cc5921d7bcc6b8850f865 Nov 27 11:00:01 crc kubenswrapper[4806]: I1127 11:00:01.478019 4806 generic.go:334] "Generic (PLEG): container finished" podID="9169c112-b10b-4938-bf73-c4f360d886ac" containerID="ec3bb98c00bb62c8a6d3dddba70abb5b23b459c27c41f4a4442317f1aaf55923" exitCode=0 Nov 27 11:00:01 crc kubenswrapper[4806]: I1127 11:00:01.478122 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" event={"ID":"9169c112-b10b-4938-bf73-c4f360d886ac","Type":"ContainerDied","Data":"ec3bb98c00bb62c8a6d3dddba70abb5b23b459c27c41f4a4442317f1aaf55923"} Nov 27 11:00:01 crc kubenswrapper[4806]: I1127 11:00:01.479732 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" event={"ID":"9169c112-b10b-4938-bf73-c4f360d886ac","Type":"ContainerStarted","Data":"73388ee0f606c4f1531f1bcf839c6569ec758c8dc19cc5921d7bcc6b8850f865"} Nov 27 11:00:02 crc kubenswrapper[4806]: I1127 11:00:02.866051 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" Nov 27 11:00:02 crc kubenswrapper[4806]: I1127 11:00:02.952159 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9169c112-b10b-4938-bf73-c4f360d886ac-config-volume\") pod \"9169c112-b10b-4938-bf73-c4f360d886ac\" (UID: \"9169c112-b10b-4938-bf73-c4f360d886ac\") " Nov 27 11:00:02 crc kubenswrapper[4806]: I1127 11:00:02.952364 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9169c112-b10b-4938-bf73-c4f360d886ac-secret-volume\") pod \"9169c112-b10b-4938-bf73-c4f360d886ac\" (UID: \"9169c112-b10b-4938-bf73-c4f360d886ac\") " Nov 27 11:00:02 crc kubenswrapper[4806]: I1127 11:00:02.953446 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkqq7\" (UniqueName: \"kubernetes.io/projected/9169c112-b10b-4938-bf73-c4f360d886ac-kube-api-access-xkqq7\") pod \"9169c112-b10b-4938-bf73-c4f360d886ac\" (UID: \"9169c112-b10b-4938-bf73-c4f360d886ac\") " Nov 27 11:00:02 crc kubenswrapper[4806]: I1127 11:00:02.953528 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9169c112-b10b-4938-bf73-c4f360d886ac-config-volume" (OuterVolumeSpecName: "config-volume") pod "9169c112-b10b-4938-bf73-c4f360d886ac" (UID: "9169c112-b10b-4938-bf73-c4f360d886ac"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 27 11:00:02 crc kubenswrapper[4806]: I1127 11:00:02.954522 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9169c112-b10b-4938-bf73-c4f360d886ac-config-volume\") on node \"crc\" DevicePath \"\"" Nov 27 11:00:02 crc kubenswrapper[4806]: I1127 11:00:02.963578 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9169c112-b10b-4938-bf73-c4f360d886ac-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9169c112-b10b-4938-bf73-c4f360d886ac" (UID: "9169c112-b10b-4938-bf73-c4f360d886ac"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:00:02 crc kubenswrapper[4806]: I1127 11:00:02.964441 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9169c112-b10b-4938-bf73-c4f360d886ac-kube-api-access-xkqq7" (OuterVolumeSpecName: "kube-api-access-xkqq7") pod "9169c112-b10b-4938-bf73-c4f360d886ac" (UID: "9169c112-b10b-4938-bf73-c4f360d886ac"). InnerVolumeSpecName "kube-api-access-xkqq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:00:03 crc kubenswrapper[4806]: I1127 11:00:03.056459 4806 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9169c112-b10b-4938-bf73-c4f360d886ac-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 27 11:00:03 crc kubenswrapper[4806]: I1127 11:00:03.056524 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkqq7\" (UniqueName: \"kubernetes.io/projected/9169c112-b10b-4938-bf73-c4f360d886ac-kube-api-access-xkqq7\") on node \"crc\" DevicePath \"\"" Nov 27 11:00:03 crc kubenswrapper[4806]: I1127 11:00:03.500189 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" event={"ID":"9169c112-b10b-4938-bf73-c4f360d886ac","Type":"ContainerDied","Data":"73388ee0f606c4f1531f1bcf839c6569ec758c8dc19cc5921d7bcc6b8850f865"} Nov 27 11:00:03 crc kubenswrapper[4806]: I1127 11:00:03.500507 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73388ee0f606c4f1531f1bcf839c6569ec758c8dc19cc5921d7bcc6b8850f865" Nov 27 11:00:03 crc kubenswrapper[4806]: I1127 11:00:03.500272 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29404020-mh5cx" Nov 27 11:00:03 crc kubenswrapper[4806]: I1127 11:00:03.968500 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk"] Nov 27 11:00:03 crc kubenswrapper[4806]: I1127 11:00:03.978342 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29403975-mz4zk"] Nov 27 11:00:04 crc kubenswrapper[4806]: I1127 11:00:04.130063 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7308dfe6-f141-4f20-a966-075d17882333" path="/var/lib/kubelet/pods/7308dfe6-f141-4f20-a966-075d17882333/volumes" Nov 27 11:00:06 crc kubenswrapper[4806]: I1127 11:00:06.130895 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:00:06 crc kubenswrapper[4806]: E1127 11:00:06.131888 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:00:20 crc kubenswrapper[4806]: I1127 11:00:20.123080 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:00:20 crc kubenswrapper[4806]: E1127 11:00:20.124311 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.708026 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vbn8q/must-gather-7wk8j"] Nov 27 11:00:23 crc kubenswrapper[4806]: E1127 11:00:23.708906 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9169c112-b10b-4938-bf73-c4f360d886ac" containerName="collect-profiles" Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.708920 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="9169c112-b10b-4938-bf73-c4f360d886ac" containerName="collect-profiles" Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.709131 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="9169c112-b10b-4938-bf73-c4f360d886ac" containerName="collect-profiles" Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.710354 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vbn8q/must-gather-7wk8j" Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.714186 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vbn8q"/"kube-root-ca.crt" Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.715122 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-vbn8q"/"default-dockercfg-8rkpk" Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.734386 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vbn8q"/"openshift-service-ca.crt" Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.746523 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vbn8q/must-gather-7wk8j"] Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.834535 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdsbh\" (UniqueName: \"kubernetes.io/projected/d3e2d05c-f6bb-4072-a353-d811c8d67307-kube-api-access-zdsbh\") pod \"must-gather-7wk8j\" (UID: \"d3e2d05c-f6bb-4072-a353-d811c8d67307\") " pod="openshift-must-gather-vbn8q/must-gather-7wk8j" Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.834785 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d3e2d05c-f6bb-4072-a353-d811c8d67307-must-gather-output\") pod \"must-gather-7wk8j\" (UID: \"d3e2d05c-f6bb-4072-a353-d811c8d67307\") " pod="openshift-must-gather-vbn8q/must-gather-7wk8j" Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.937007 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdsbh\" (UniqueName: \"kubernetes.io/projected/d3e2d05c-f6bb-4072-a353-d811c8d67307-kube-api-access-zdsbh\") pod \"must-gather-7wk8j\" (UID: \"d3e2d05c-f6bb-4072-a353-d811c8d67307\") " pod="openshift-must-gather-vbn8q/must-gather-7wk8j" Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.937075 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d3e2d05c-f6bb-4072-a353-d811c8d67307-must-gather-output\") pod \"must-gather-7wk8j\" (UID: \"d3e2d05c-f6bb-4072-a353-d811c8d67307\") " pod="openshift-must-gather-vbn8q/must-gather-7wk8j" Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.937497 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d3e2d05c-f6bb-4072-a353-d811c8d67307-must-gather-output\") pod \"must-gather-7wk8j\" (UID: \"d3e2d05c-f6bb-4072-a353-d811c8d67307\") " pod="openshift-must-gather-vbn8q/must-gather-7wk8j" Nov 27 11:00:23 crc kubenswrapper[4806]: I1127 11:00:23.962058 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdsbh\" (UniqueName: \"kubernetes.io/projected/d3e2d05c-f6bb-4072-a353-d811c8d67307-kube-api-access-zdsbh\") pod \"must-gather-7wk8j\" (UID: \"d3e2d05c-f6bb-4072-a353-d811c8d67307\") " pod="openshift-must-gather-vbn8q/must-gather-7wk8j" Nov 27 11:00:24 crc kubenswrapper[4806]: I1127 11:00:24.035333 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vbn8q/must-gather-7wk8j" Nov 27 11:00:24 crc kubenswrapper[4806]: I1127 11:00:24.604928 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vbn8q/must-gather-7wk8j"] Nov 27 11:00:24 crc kubenswrapper[4806]: I1127 11:00:24.702346 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vbn8q/must-gather-7wk8j" event={"ID":"d3e2d05c-f6bb-4072-a353-d811c8d67307","Type":"ContainerStarted","Data":"a3a8bd76c82a55e55655afbdfc3cfabac34fc4a0ec026cad940520da832d2a8e"} Nov 27 11:00:32 crc kubenswrapper[4806]: I1127 11:00:32.121848 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:00:32 crc kubenswrapper[4806]: E1127 11:00:32.122635 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:00:33 crc kubenswrapper[4806]: I1127 11:00:33.851663 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vbn8q/must-gather-7wk8j" event={"ID":"d3e2d05c-f6bb-4072-a353-d811c8d67307","Type":"ContainerStarted","Data":"fa261d446bd2845eb5d06e7e61b4ad539b9c5b0ecf95406bdda373790e8c4475"} Nov 27 11:00:34 crc kubenswrapper[4806]: I1127 11:00:34.862905 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vbn8q/must-gather-7wk8j" event={"ID":"d3e2d05c-f6bb-4072-a353-d811c8d67307","Type":"ContainerStarted","Data":"40154098684ec914750b51d528275ceda1c92aec958d89187daa648e688e5f89"} Nov 27 11:00:34 crc kubenswrapper[4806]: I1127 11:00:34.885490 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vbn8q/must-gather-7wk8j" podStartSLOduration=2.902720463 podStartE2EDuration="11.88547048s" podCreationTimestamp="2025-11-27 11:00:23 +0000 UTC" firstStartedPulling="2025-11-27 11:00:24.608473178 +0000 UTC m=+2329.195063942" lastFinishedPulling="2025-11-27 11:00:33.591223195 +0000 UTC m=+2338.177813959" observedRunningTime="2025-11-27 11:00:34.883961448 +0000 UTC m=+2339.470552242" watchObservedRunningTime="2025-11-27 11:00:34.88547048 +0000 UTC m=+2339.472061234" Nov 27 11:00:38 crc kubenswrapper[4806]: I1127 11:00:38.379060 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vbn8q/crc-debug-z859l"] Nov 27 11:00:38 crc kubenswrapper[4806]: I1127 11:00:38.381079 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vbn8q/crc-debug-z859l" Nov 27 11:00:38 crc kubenswrapper[4806]: I1127 11:00:38.514480 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/18cf012b-ba51-4d6b-9227-a41ba7980f3f-host\") pod \"crc-debug-z859l\" (UID: \"18cf012b-ba51-4d6b-9227-a41ba7980f3f\") " pod="openshift-must-gather-vbn8q/crc-debug-z859l" Nov 27 11:00:38 crc kubenswrapper[4806]: I1127 11:00:38.514626 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfll9\" (UniqueName: \"kubernetes.io/projected/18cf012b-ba51-4d6b-9227-a41ba7980f3f-kube-api-access-sfll9\") pod \"crc-debug-z859l\" (UID: \"18cf012b-ba51-4d6b-9227-a41ba7980f3f\") " pod="openshift-must-gather-vbn8q/crc-debug-z859l" Nov 27 11:00:38 crc kubenswrapper[4806]: I1127 11:00:38.617281 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/18cf012b-ba51-4d6b-9227-a41ba7980f3f-host\") pod \"crc-debug-z859l\" (UID: \"18cf012b-ba51-4d6b-9227-a41ba7980f3f\") " pod="openshift-must-gather-vbn8q/crc-debug-z859l" Nov 27 11:00:38 crc kubenswrapper[4806]: I1127 11:00:38.617411 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfll9\" (UniqueName: \"kubernetes.io/projected/18cf012b-ba51-4d6b-9227-a41ba7980f3f-kube-api-access-sfll9\") pod \"crc-debug-z859l\" (UID: \"18cf012b-ba51-4d6b-9227-a41ba7980f3f\") " pod="openshift-must-gather-vbn8q/crc-debug-z859l" Nov 27 11:00:38 crc kubenswrapper[4806]: I1127 11:00:38.617961 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/18cf012b-ba51-4d6b-9227-a41ba7980f3f-host\") pod \"crc-debug-z859l\" (UID: \"18cf012b-ba51-4d6b-9227-a41ba7980f3f\") " pod="openshift-must-gather-vbn8q/crc-debug-z859l" Nov 27 11:00:38 crc kubenswrapper[4806]: I1127 11:00:38.641322 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfll9\" (UniqueName: \"kubernetes.io/projected/18cf012b-ba51-4d6b-9227-a41ba7980f3f-kube-api-access-sfll9\") pod \"crc-debug-z859l\" (UID: \"18cf012b-ba51-4d6b-9227-a41ba7980f3f\") " pod="openshift-must-gather-vbn8q/crc-debug-z859l" Nov 27 11:00:38 crc kubenswrapper[4806]: I1127 11:00:38.712259 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vbn8q/crc-debug-z859l" Nov 27 11:00:38 crc kubenswrapper[4806]: W1127 11:00:38.750155 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18cf012b_ba51_4d6b_9227_a41ba7980f3f.slice/crio-ca670a227cbd037ec06b7b01716a2a11b71c9d4543314df6bf4483543bd1722e WatchSource:0}: Error finding container ca670a227cbd037ec06b7b01716a2a11b71c9d4543314df6bf4483543bd1722e: Status 404 returned error can't find the container with id ca670a227cbd037ec06b7b01716a2a11b71c9d4543314df6bf4483543bd1722e Nov 27 11:00:38 crc kubenswrapper[4806]: I1127 11:00:38.912404 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vbn8q/crc-debug-z859l" event={"ID":"18cf012b-ba51-4d6b-9227-a41ba7980f3f","Type":"ContainerStarted","Data":"ca670a227cbd037ec06b7b01716a2a11b71c9d4543314df6bf4483543bd1722e"} Nov 27 11:00:44 crc kubenswrapper[4806]: I1127 11:00:44.118163 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:00:44 crc kubenswrapper[4806]: E1127 11:00:44.119563 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:00:47 crc kubenswrapper[4806]: I1127 11:00:47.008306 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dn5ql"] Nov 27 11:00:47 crc kubenswrapper[4806]: I1127 11:00:47.012201 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:00:47 crc kubenswrapper[4806]: I1127 11:00:47.047350 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dn5ql"] Nov 27 11:00:47 crc kubenswrapper[4806]: I1127 11:00:47.112266 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4d26cb5-3562-46a5-87da-8561886fe273-catalog-content\") pod \"redhat-operators-dn5ql\" (UID: \"d4d26cb5-3562-46a5-87da-8561886fe273\") " pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:00:47 crc kubenswrapper[4806]: I1127 11:00:47.112419 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4d26cb5-3562-46a5-87da-8561886fe273-utilities\") pod \"redhat-operators-dn5ql\" (UID: \"d4d26cb5-3562-46a5-87da-8561886fe273\") " pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:00:47 crc kubenswrapper[4806]: I1127 11:00:47.112477 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd5jd\" (UniqueName: \"kubernetes.io/projected/d4d26cb5-3562-46a5-87da-8561886fe273-kube-api-access-sd5jd\") pod \"redhat-operators-dn5ql\" (UID: \"d4d26cb5-3562-46a5-87da-8561886fe273\") " pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:00:47 crc kubenswrapper[4806]: I1127 11:00:47.214851 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4d26cb5-3562-46a5-87da-8561886fe273-catalog-content\") pod \"redhat-operators-dn5ql\" (UID: \"d4d26cb5-3562-46a5-87da-8561886fe273\") " pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:00:47 crc kubenswrapper[4806]: I1127 11:00:47.214931 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4d26cb5-3562-46a5-87da-8561886fe273-utilities\") pod \"redhat-operators-dn5ql\" (UID: \"d4d26cb5-3562-46a5-87da-8561886fe273\") " pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:00:47 crc kubenswrapper[4806]: I1127 11:00:47.214969 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd5jd\" (UniqueName: \"kubernetes.io/projected/d4d26cb5-3562-46a5-87da-8561886fe273-kube-api-access-sd5jd\") pod \"redhat-operators-dn5ql\" (UID: \"d4d26cb5-3562-46a5-87da-8561886fe273\") " pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:00:47 crc kubenswrapper[4806]: I1127 11:00:47.215953 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4d26cb5-3562-46a5-87da-8561886fe273-catalog-content\") pod \"redhat-operators-dn5ql\" (UID: \"d4d26cb5-3562-46a5-87da-8561886fe273\") " pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:00:47 crc kubenswrapper[4806]: I1127 11:00:47.216336 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4d26cb5-3562-46a5-87da-8561886fe273-utilities\") pod \"redhat-operators-dn5ql\" (UID: \"d4d26cb5-3562-46a5-87da-8561886fe273\") " pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:00:47 crc kubenswrapper[4806]: I1127 11:00:47.239047 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd5jd\" (UniqueName: \"kubernetes.io/projected/d4d26cb5-3562-46a5-87da-8561886fe273-kube-api-access-sd5jd\") pod \"redhat-operators-dn5ql\" (UID: \"d4d26cb5-3562-46a5-87da-8561886fe273\") " pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:00:47 crc kubenswrapper[4806]: I1127 11:00:47.352335 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:00:48 crc kubenswrapper[4806]: I1127 11:00:48.040318 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dn5ql"] Nov 27 11:00:49 crc kubenswrapper[4806]: I1127 11:00:49.077726 4806 generic.go:334] "Generic (PLEG): container finished" podID="d4d26cb5-3562-46a5-87da-8561886fe273" containerID="964af62703746475aa6be08ae0657b8f5cbc38d96ba79fa76b5248951906238c" exitCode=0 Nov 27 11:00:49 crc kubenswrapper[4806]: I1127 11:00:49.077829 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dn5ql" event={"ID":"d4d26cb5-3562-46a5-87da-8561886fe273","Type":"ContainerDied","Data":"964af62703746475aa6be08ae0657b8f5cbc38d96ba79fa76b5248951906238c"} Nov 27 11:00:49 crc kubenswrapper[4806]: I1127 11:00:49.078145 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dn5ql" event={"ID":"d4d26cb5-3562-46a5-87da-8561886fe273","Type":"ContainerStarted","Data":"d27e744a9f4bddc48c8263ee8da6e0966cd05333de6c62633077325fbe4cf4c6"} Nov 27 11:00:50 crc kubenswrapper[4806]: I1127 11:00:50.088001 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dn5ql" event={"ID":"d4d26cb5-3562-46a5-87da-8561886fe273","Type":"ContainerStarted","Data":"0539a73bf12da02292d191b3f92a43a52330623ca4798e15b59ea785a6392188"} Nov 27 11:00:54 crc kubenswrapper[4806]: I1127 11:00:54.508614 4806 scope.go:117] "RemoveContainer" containerID="535ec0a3c11c7d698bab89bdd18508c2a1ab57d28df16f3cf0b6cd72803ce638" Nov 27 11:00:59 crc kubenswrapper[4806]: I1127 11:00:59.117902 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:00:59 crc kubenswrapper[4806]: E1127 11:00:59.119724 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:00:59 crc kubenswrapper[4806]: E1127 11:00:59.567163 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Nov 27 11:00:59 crc kubenswrapper[4806]: E1127 11:00:59.567397 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sfll9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-z859l_openshift-must-gather-vbn8q(18cf012b-ba51-4d6b-9227-a41ba7980f3f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 27 11:00:59 crc kubenswrapper[4806]: E1127 11:00:59.568589 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-vbn8q/crc-debug-z859l" podUID="18cf012b-ba51-4d6b-9227-a41ba7980f3f" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.186264 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29404021-wxzgw"] Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.187918 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.204574 4806 generic.go:334] "Generic (PLEG): container finished" podID="d4d26cb5-3562-46a5-87da-8561886fe273" containerID="0539a73bf12da02292d191b3f92a43a52330623ca4798e15b59ea785a6392188" exitCode=0 Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.204731 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dn5ql" event={"ID":"d4d26cb5-3562-46a5-87da-8561886fe273","Type":"ContainerDied","Data":"0539a73bf12da02292d191b3f92a43a52330623ca4798e15b59ea785a6392188"} Nov 27 11:01:00 crc kubenswrapper[4806]: E1127 11:01:00.206912 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-vbn8q/crc-debug-z859l" podUID="18cf012b-ba51-4d6b-9227-a41ba7980f3f" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.207963 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.218213 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29404021-wxzgw"] Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.257981 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-config-data\") pod \"keystone-cron-29404021-wxzgw\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.258072 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5vjq\" (UniqueName: \"kubernetes.io/projected/9c0737de-4b3d-42ba-8a59-068721ee9828-kube-api-access-n5vjq\") pod \"keystone-cron-29404021-wxzgw\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.258100 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-fernet-keys\") pod \"keystone-cron-29404021-wxzgw\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.258169 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-combined-ca-bundle\") pod \"keystone-cron-29404021-wxzgw\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.360469 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-combined-ca-bundle\") pod \"keystone-cron-29404021-wxzgw\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.360629 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-config-data\") pod \"keystone-cron-29404021-wxzgw\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.360673 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5vjq\" (UniqueName: \"kubernetes.io/projected/9c0737de-4b3d-42ba-8a59-068721ee9828-kube-api-access-n5vjq\") pod \"keystone-cron-29404021-wxzgw\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.360697 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-fernet-keys\") pod \"keystone-cron-29404021-wxzgw\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.367847 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-config-data\") pod \"keystone-cron-29404021-wxzgw\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.370038 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-fernet-keys\") pod \"keystone-cron-29404021-wxzgw\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.397309 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5vjq\" (UniqueName: \"kubernetes.io/projected/9c0737de-4b3d-42ba-8a59-068721ee9828-kube-api-access-n5vjq\") pod \"keystone-cron-29404021-wxzgw\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.405375 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-combined-ca-bundle\") pod \"keystone-cron-29404021-wxzgw\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:00 crc kubenswrapper[4806]: I1127 11:01:00.509495 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:01 crc kubenswrapper[4806]: I1127 11:01:01.234196 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29404021-wxzgw"] Nov 27 11:01:02 crc kubenswrapper[4806]: I1127 11:01:02.248367 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dn5ql" event={"ID":"d4d26cb5-3562-46a5-87da-8561886fe273","Type":"ContainerStarted","Data":"d53caf74dd1ec00020dd4bbe5ba2e34e0ada4918fbb7f286e7308f021737757e"} Nov 27 11:01:02 crc kubenswrapper[4806]: I1127 11:01:02.252327 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29404021-wxzgw" event={"ID":"9c0737de-4b3d-42ba-8a59-068721ee9828","Type":"ContainerStarted","Data":"8db8b5d83c30e8b880d6d09d7e2ee2a588f1ec0d27eca60ad49abb75595f0f6a"} Nov 27 11:01:02 crc kubenswrapper[4806]: I1127 11:01:02.252365 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29404021-wxzgw" event={"ID":"9c0737de-4b3d-42ba-8a59-068721ee9828","Type":"ContainerStarted","Data":"4f88c8c7186d5a4fd9e8871b4713a4a02c46192dcf01a540cc7386987d484a56"} Nov 27 11:01:02 crc kubenswrapper[4806]: I1127 11:01:02.311715 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dn5ql" podStartSLOduration=4.121817113 podStartE2EDuration="16.311694167s" podCreationTimestamp="2025-11-27 11:00:46 +0000 UTC" firstStartedPulling="2025-11-27 11:00:49.081425372 +0000 UTC m=+2353.668016136" lastFinishedPulling="2025-11-27 11:01:01.271302426 +0000 UTC m=+2365.857893190" observedRunningTime="2025-11-27 11:01:02.304883609 +0000 UTC m=+2366.891474373" watchObservedRunningTime="2025-11-27 11:01:02.311694167 +0000 UTC m=+2366.898284931" Nov 27 11:01:02 crc kubenswrapper[4806]: I1127 11:01:02.352384 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29404021-wxzgw" podStartSLOduration=2.352360189 podStartE2EDuration="2.352360189s" podCreationTimestamp="2025-11-27 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-27 11:01:02.350913159 +0000 UTC m=+2366.937503923" watchObservedRunningTime="2025-11-27 11:01:02.352360189 +0000 UTC m=+2366.938950953" Nov 27 11:01:05 crc kubenswrapper[4806]: I1127 11:01:05.290838 4806 generic.go:334] "Generic (PLEG): container finished" podID="9c0737de-4b3d-42ba-8a59-068721ee9828" containerID="8db8b5d83c30e8b880d6d09d7e2ee2a588f1ec0d27eca60ad49abb75595f0f6a" exitCode=0 Nov 27 11:01:05 crc kubenswrapper[4806]: I1127 11:01:05.290903 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29404021-wxzgw" event={"ID":"9c0737de-4b3d-42ba-8a59-068721ee9828","Type":"ContainerDied","Data":"8db8b5d83c30e8b880d6d09d7e2ee2a588f1ec0d27eca60ad49abb75595f0f6a"} Nov 27 11:01:06 crc kubenswrapper[4806]: I1127 11:01:06.889874 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:06 crc kubenswrapper[4806]: I1127 11:01:06.926142 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-config-data\") pod \"9c0737de-4b3d-42ba-8a59-068721ee9828\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " Nov 27 11:01:06 crc kubenswrapper[4806]: I1127 11:01:06.926579 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5vjq\" (UniqueName: \"kubernetes.io/projected/9c0737de-4b3d-42ba-8a59-068721ee9828-kube-api-access-n5vjq\") pod \"9c0737de-4b3d-42ba-8a59-068721ee9828\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " Nov 27 11:01:06 crc kubenswrapper[4806]: I1127 11:01:06.926729 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-fernet-keys\") pod \"9c0737de-4b3d-42ba-8a59-068721ee9828\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " Nov 27 11:01:06 crc kubenswrapper[4806]: I1127 11:01:06.927031 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-combined-ca-bundle\") pod \"9c0737de-4b3d-42ba-8a59-068721ee9828\" (UID: \"9c0737de-4b3d-42ba-8a59-068721ee9828\") " Nov 27 11:01:06 crc kubenswrapper[4806]: I1127 11:01:06.963687 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9c0737de-4b3d-42ba-8a59-068721ee9828" (UID: "9c0737de-4b3d-42ba-8a59-068721ee9828"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:01:06 crc kubenswrapper[4806]: I1127 11:01:06.963907 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c0737de-4b3d-42ba-8a59-068721ee9828-kube-api-access-n5vjq" (OuterVolumeSpecName: "kube-api-access-n5vjq") pod "9c0737de-4b3d-42ba-8a59-068721ee9828" (UID: "9c0737de-4b3d-42ba-8a59-068721ee9828"). InnerVolumeSpecName "kube-api-access-n5vjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:01:06 crc kubenswrapper[4806]: I1127 11:01:06.967586 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c0737de-4b3d-42ba-8a59-068721ee9828" (UID: "9c0737de-4b3d-42ba-8a59-068721ee9828"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:01:07 crc kubenswrapper[4806]: I1127 11:01:07.030165 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 27 11:01:07 crc kubenswrapper[4806]: I1127 11:01:07.030464 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5vjq\" (UniqueName: \"kubernetes.io/projected/9c0737de-4b3d-42ba-8a59-068721ee9828-kube-api-access-n5vjq\") on node \"crc\" DevicePath \"\"" Nov 27 11:01:07 crc kubenswrapper[4806]: I1127 11:01:07.030480 4806 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 27 11:01:07 crc kubenswrapper[4806]: I1127 11:01:07.074892 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-config-data" (OuterVolumeSpecName: "config-data") pod "9c0737de-4b3d-42ba-8a59-068721ee9828" (UID: "9c0737de-4b3d-42ba-8a59-068721ee9828"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 27 11:01:07 crc kubenswrapper[4806]: I1127 11:01:07.132436 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c0737de-4b3d-42ba-8a59-068721ee9828-config-data\") on node \"crc\" DevicePath \"\"" Nov 27 11:01:07 crc kubenswrapper[4806]: I1127 11:01:07.310447 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29404021-wxzgw" event={"ID":"9c0737de-4b3d-42ba-8a59-068721ee9828","Type":"ContainerDied","Data":"4f88c8c7186d5a4fd9e8871b4713a4a02c46192dcf01a540cc7386987d484a56"} Nov 27 11:01:07 crc kubenswrapper[4806]: I1127 11:01:07.310494 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f88c8c7186d5a4fd9e8871b4713a4a02c46192dcf01a540cc7386987d484a56" Nov 27 11:01:07 crc kubenswrapper[4806]: I1127 11:01:07.310526 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29404021-wxzgw" Nov 27 11:01:07 crc kubenswrapper[4806]: I1127 11:01:07.353741 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:01:07 crc kubenswrapper[4806]: I1127 11:01:07.353807 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:01:08 crc kubenswrapper[4806]: I1127 11:01:08.408954 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dn5ql" podUID="d4d26cb5-3562-46a5-87da-8561886fe273" containerName="registry-server" probeResult="failure" output=< Nov 27 11:01:08 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 11:01:08 crc kubenswrapper[4806]: > Nov 27 11:01:11 crc kubenswrapper[4806]: I1127 11:01:11.117318 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:01:11 crc kubenswrapper[4806]: E1127 11:01:11.117673 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:01:12 crc kubenswrapper[4806]: I1127 11:01:12.357324 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vbn8q/crc-debug-z859l" event={"ID":"18cf012b-ba51-4d6b-9227-a41ba7980f3f","Type":"ContainerStarted","Data":"78454c0b4d9baf62304b533edc641837163a05c699cb797517956259b2c7118e"} Nov 27 11:01:12 crc kubenswrapper[4806]: I1127 11:01:12.381840 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vbn8q/crc-debug-z859l" podStartSLOduration=1.340941823 podStartE2EDuration="34.381816951s" podCreationTimestamp="2025-11-27 11:00:38 +0000 UTC" firstStartedPulling="2025-11-27 11:00:38.753482917 +0000 UTC m=+2343.340073691" lastFinishedPulling="2025-11-27 11:01:11.794358055 +0000 UTC m=+2376.380948819" observedRunningTime="2025-11-27 11:01:12.373678156 +0000 UTC m=+2376.960268940" watchObservedRunningTime="2025-11-27 11:01:12.381816951 +0000 UTC m=+2376.968407715" Nov 27 11:01:18 crc kubenswrapper[4806]: I1127 11:01:18.411433 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dn5ql" podUID="d4d26cb5-3562-46a5-87da-8561886fe273" containerName="registry-server" probeResult="failure" output=< Nov 27 11:01:18 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 11:01:18 crc kubenswrapper[4806]: > Nov 27 11:01:26 crc kubenswrapper[4806]: I1127 11:01:26.128065 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:01:26 crc kubenswrapper[4806]: E1127 11:01:26.129276 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:01:28 crc kubenswrapper[4806]: I1127 11:01:28.412835 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dn5ql" podUID="d4d26cb5-3562-46a5-87da-8561886fe273" containerName="registry-server" probeResult="failure" output=< Nov 27 11:01:28 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 11:01:28 crc kubenswrapper[4806]: > Nov 27 11:01:30 crc kubenswrapper[4806]: I1127 11:01:30.543390 4806 generic.go:334] "Generic (PLEG): container finished" podID="18cf012b-ba51-4d6b-9227-a41ba7980f3f" containerID="78454c0b4d9baf62304b533edc641837163a05c699cb797517956259b2c7118e" exitCode=0 Nov 27 11:01:30 crc kubenswrapper[4806]: I1127 11:01:30.543469 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vbn8q/crc-debug-z859l" event={"ID":"18cf012b-ba51-4d6b-9227-a41ba7980f3f","Type":"ContainerDied","Data":"78454c0b4d9baf62304b533edc641837163a05c699cb797517956259b2c7118e"} Nov 27 11:01:31 crc kubenswrapper[4806]: I1127 11:01:31.685885 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vbn8q/crc-debug-z859l" Nov 27 11:01:31 crc kubenswrapper[4806]: I1127 11:01:31.734567 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vbn8q/crc-debug-z859l"] Nov 27 11:01:31 crc kubenswrapper[4806]: I1127 11:01:31.745349 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vbn8q/crc-debug-z859l"] Nov 27 11:01:31 crc kubenswrapper[4806]: I1127 11:01:31.772793 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/18cf012b-ba51-4d6b-9227-a41ba7980f3f-host\") pod \"18cf012b-ba51-4d6b-9227-a41ba7980f3f\" (UID: \"18cf012b-ba51-4d6b-9227-a41ba7980f3f\") " Nov 27 11:01:31 crc kubenswrapper[4806]: I1127 11:01:31.772977 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfll9\" (UniqueName: \"kubernetes.io/projected/18cf012b-ba51-4d6b-9227-a41ba7980f3f-kube-api-access-sfll9\") pod \"18cf012b-ba51-4d6b-9227-a41ba7980f3f\" (UID: \"18cf012b-ba51-4d6b-9227-a41ba7980f3f\") " Nov 27 11:01:31 crc kubenswrapper[4806]: I1127 11:01:31.772998 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/18cf012b-ba51-4d6b-9227-a41ba7980f3f-host" (OuterVolumeSpecName: "host") pod "18cf012b-ba51-4d6b-9227-a41ba7980f3f" (UID: "18cf012b-ba51-4d6b-9227-a41ba7980f3f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:01:31 crc kubenswrapper[4806]: I1127 11:01:31.773703 4806 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/18cf012b-ba51-4d6b-9227-a41ba7980f3f-host\") on node \"crc\" DevicePath \"\"" Nov 27 11:01:31 crc kubenswrapper[4806]: I1127 11:01:31.794678 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18cf012b-ba51-4d6b-9227-a41ba7980f3f-kube-api-access-sfll9" (OuterVolumeSpecName: "kube-api-access-sfll9") pod "18cf012b-ba51-4d6b-9227-a41ba7980f3f" (UID: "18cf012b-ba51-4d6b-9227-a41ba7980f3f"). InnerVolumeSpecName "kube-api-access-sfll9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:01:31 crc kubenswrapper[4806]: I1127 11:01:31.876003 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfll9\" (UniqueName: \"kubernetes.io/projected/18cf012b-ba51-4d6b-9227-a41ba7980f3f-kube-api-access-sfll9\") on node \"crc\" DevicePath \"\"" Nov 27 11:01:32 crc kubenswrapper[4806]: I1127 11:01:32.127509 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18cf012b-ba51-4d6b-9227-a41ba7980f3f" path="/var/lib/kubelet/pods/18cf012b-ba51-4d6b-9227-a41ba7980f3f/volumes" Nov 27 11:01:32 crc kubenswrapper[4806]: I1127 11:01:32.563707 4806 scope.go:117] "RemoveContainer" containerID="78454c0b4d9baf62304b533edc641837163a05c699cb797517956259b2c7118e" Nov 27 11:01:32 crc kubenswrapper[4806]: I1127 11:01:32.563765 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vbn8q/crc-debug-z859l" Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.113260 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vbn8q/crc-debug-7l5zf"] Nov 27 11:01:33 crc kubenswrapper[4806]: E1127 11:01:33.114280 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18cf012b-ba51-4d6b-9227-a41ba7980f3f" containerName="container-00" Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.114301 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="18cf012b-ba51-4d6b-9227-a41ba7980f3f" containerName="container-00" Nov 27 11:01:33 crc kubenswrapper[4806]: E1127 11:01:33.114353 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c0737de-4b3d-42ba-8a59-068721ee9828" containerName="keystone-cron" Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.114362 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c0737de-4b3d-42ba-8a59-068721ee9828" containerName="keystone-cron" Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.114582 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="18cf012b-ba51-4d6b-9227-a41ba7980f3f" containerName="container-00" Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.114613 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c0737de-4b3d-42ba-8a59-068721ee9828" containerName="keystone-cron" Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.115509 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vbn8q/crc-debug-7l5zf" Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.213489 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/66620702-1b55-49ef-babc-4d43a8767274-host\") pod \"crc-debug-7l5zf\" (UID: \"66620702-1b55-49ef-babc-4d43a8767274\") " pod="openshift-must-gather-vbn8q/crc-debug-7l5zf" Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.213546 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mld7x\" (UniqueName: \"kubernetes.io/projected/66620702-1b55-49ef-babc-4d43a8767274-kube-api-access-mld7x\") pod \"crc-debug-7l5zf\" (UID: \"66620702-1b55-49ef-babc-4d43a8767274\") " pod="openshift-must-gather-vbn8q/crc-debug-7l5zf" Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.316080 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/66620702-1b55-49ef-babc-4d43a8767274-host\") pod \"crc-debug-7l5zf\" (UID: \"66620702-1b55-49ef-babc-4d43a8767274\") " pod="openshift-must-gather-vbn8q/crc-debug-7l5zf" Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.316157 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mld7x\" (UniqueName: \"kubernetes.io/projected/66620702-1b55-49ef-babc-4d43a8767274-kube-api-access-mld7x\") pod \"crc-debug-7l5zf\" (UID: \"66620702-1b55-49ef-babc-4d43a8767274\") " pod="openshift-must-gather-vbn8q/crc-debug-7l5zf" Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.316276 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/66620702-1b55-49ef-babc-4d43a8767274-host\") pod \"crc-debug-7l5zf\" (UID: \"66620702-1b55-49ef-babc-4d43a8767274\") " pod="openshift-must-gather-vbn8q/crc-debug-7l5zf" Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.353747 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mld7x\" (UniqueName: \"kubernetes.io/projected/66620702-1b55-49ef-babc-4d43a8767274-kube-api-access-mld7x\") pod \"crc-debug-7l5zf\" (UID: \"66620702-1b55-49ef-babc-4d43a8767274\") " pod="openshift-must-gather-vbn8q/crc-debug-7l5zf" Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.463043 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vbn8q/crc-debug-7l5zf" Nov 27 11:01:33 crc kubenswrapper[4806]: W1127 11:01:33.515600 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66620702_1b55_49ef_babc_4d43a8767274.slice/crio-594a309d96f6bb694c96e3bbf833f854736bb7aa528058175cc3dca67630545c WatchSource:0}: Error finding container 594a309d96f6bb694c96e3bbf833f854736bb7aa528058175cc3dca67630545c: Status 404 returned error can't find the container with id 594a309d96f6bb694c96e3bbf833f854736bb7aa528058175cc3dca67630545c Nov 27 11:01:33 crc kubenswrapper[4806]: I1127 11:01:33.576106 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vbn8q/crc-debug-7l5zf" event={"ID":"66620702-1b55-49ef-babc-4d43a8767274","Type":"ContainerStarted","Data":"594a309d96f6bb694c96e3bbf833f854736bb7aa528058175cc3dca67630545c"} Nov 27 11:01:34 crc kubenswrapper[4806]: I1127 11:01:34.594818 4806 generic.go:334] "Generic (PLEG): container finished" podID="66620702-1b55-49ef-babc-4d43a8767274" containerID="ca58034f4cbe9744985f3913a5693edc8ffa3bf4d621f8c8f6e04de206ebe4c8" exitCode=1 Nov 27 11:01:34 crc kubenswrapper[4806]: I1127 11:01:34.595331 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vbn8q/crc-debug-7l5zf" event={"ID":"66620702-1b55-49ef-babc-4d43a8767274","Type":"ContainerDied","Data":"ca58034f4cbe9744985f3913a5693edc8ffa3bf4d621f8c8f6e04de206ebe4c8"} Nov 27 11:01:34 crc kubenswrapper[4806]: I1127 11:01:34.638409 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vbn8q/crc-debug-7l5zf"] Nov 27 11:01:34 crc kubenswrapper[4806]: I1127 11:01:34.648080 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vbn8q/crc-debug-7l5zf"] Nov 27 11:01:35 crc kubenswrapper[4806]: I1127 11:01:35.738189 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vbn8q/crc-debug-7l5zf" Nov 27 11:01:35 crc kubenswrapper[4806]: I1127 11:01:35.772416 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/66620702-1b55-49ef-babc-4d43a8767274-host\") pod \"66620702-1b55-49ef-babc-4d43a8767274\" (UID: \"66620702-1b55-49ef-babc-4d43a8767274\") " Nov 27 11:01:35 crc kubenswrapper[4806]: I1127 11:01:35.772630 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66620702-1b55-49ef-babc-4d43a8767274-host" (OuterVolumeSpecName: "host") pod "66620702-1b55-49ef-babc-4d43a8767274" (UID: "66620702-1b55-49ef-babc-4d43a8767274"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 27 11:01:35 crc kubenswrapper[4806]: I1127 11:01:35.772892 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mld7x\" (UniqueName: \"kubernetes.io/projected/66620702-1b55-49ef-babc-4d43a8767274-kube-api-access-mld7x\") pod \"66620702-1b55-49ef-babc-4d43a8767274\" (UID: \"66620702-1b55-49ef-babc-4d43a8767274\") " Nov 27 11:01:35 crc kubenswrapper[4806]: I1127 11:01:35.773618 4806 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/66620702-1b55-49ef-babc-4d43a8767274-host\") on node \"crc\" DevicePath \"\"" Nov 27 11:01:35 crc kubenswrapper[4806]: I1127 11:01:35.786665 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66620702-1b55-49ef-babc-4d43a8767274-kube-api-access-mld7x" (OuterVolumeSpecName: "kube-api-access-mld7x") pod "66620702-1b55-49ef-babc-4d43a8767274" (UID: "66620702-1b55-49ef-babc-4d43a8767274"). InnerVolumeSpecName "kube-api-access-mld7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:01:35 crc kubenswrapper[4806]: I1127 11:01:35.879530 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mld7x\" (UniqueName: \"kubernetes.io/projected/66620702-1b55-49ef-babc-4d43a8767274-kube-api-access-mld7x\") on node \"crc\" DevicePath \"\"" Nov 27 11:01:36 crc kubenswrapper[4806]: I1127 11:01:36.139540 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66620702-1b55-49ef-babc-4d43a8767274" path="/var/lib/kubelet/pods/66620702-1b55-49ef-babc-4d43a8767274/volumes" Nov 27 11:01:36 crc kubenswrapper[4806]: I1127 11:01:36.623921 4806 scope.go:117] "RemoveContainer" containerID="ca58034f4cbe9744985f3913a5693edc8ffa3bf4d621f8c8f6e04de206ebe4c8" Nov 27 11:01:36 crc kubenswrapper[4806]: I1127 11:01:36.624417 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vbn8q/crc-debug-7l5zf" Nov 27 11:01:37 crc kubenswrapper[4806]: I1127 11:01:37.425874 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:01:37 crc kubenswrapper[4806]: I1127 11:01:37.488884 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:01:40 crc kubenswrapper[4806]: I1127 11:01:40.117053 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:01:40 crc kubenswrapper[4806]: E1127 11:01:40.117378 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:01:40 crc kubenswrapper[4806]: I1127 11:01:40.406712 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dn5ql"] Nov 27 11:01:40 crc kubenswrapper[4806]: I1127 11:01:40.407023 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dn5ql" podUID="d4d26cb5-3562-46a5-87da-8561886fe273" containerName="registry-server" containerID="cri-o://d53caf74dd1ec00020dd4bbe5ba2e34e0ada4918fbb7f286e7308f021737757e" gracePeriod=2 Nov 27 11:01:40 crc kubenswrapper[4806]: I1127 11:01:40.671532 4806 generic.go:334] "Generic (PLEG): container finished" podID="d4d26cb5-3562-46a5-87da-8561886fe273" containerID="d53caf74dd1ec00020dd4bbe5ba2e34e0ada4918fbb7f286e7308f021737757e" exitCode=0 Nov 27 11:01:40 crc kubenswrapper[4806]: I1127 11:01:40.671979 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dn5ql" event={"ID":"d4d26cb5-3562-46a5-87da-8561886fe273","Type":"ContainerDied","Data":"d53caf74dd1ec00020dd4bbe5ba2e34e0ada4918fbb7f286e7308f021737757e"} Nov 27 11:01:40 crc kubenswrapper[4806]: I1127 11:01:40.911828 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.002130 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sd5jd\" (UniqueName: \"kubernetes.io/projected/d4d26cb5-3562-46a5-87da-8561886fe273-kube-api-access-sd5jd\") pod \"d4d26cb5-3562-46a5-87da-8561886fe273\" (UID: \"d4d26cb5-3562-46a5-87da-8561886fe273\") " Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.002371 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4d26cb5-3562-46a5-87da-8561886fe273-utilities\") pod \"d4d26cb5-3562-46a5-87da-8561886fe273\" (UID: \"d4d26cb5-3562-46a5-87da-8561886fe273\") " Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.002511 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4d26cb5-3562-46a5-87da-8561886fe273-catalog-content\") pod \"d4d26cb5-3562-46a5-87da-8561886fe273\" (UID: \"d4d26cb5-3562-46a5-87da-8561886fe273\") " Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.004977 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4d26cb5-3562-46a5-87da-8561886fe273-utilities" (OuterVolumeSpecName: "utilities") pod "d4d26cb5-3562-46a5-87da-8561886fe273" (UID: "d4d26cb5-3562-46a5-87da-8561886fe273"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.042445 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4d26cb5-3562-46a5-87da-8561886fe273-kube-api-access-sd5jd" (OuterVolumeSpecName: "kube-api-access-sd5jd") pod "d4d26cb5-3562-46a5-87da-8561886fe273" (UID: "d4d26cb5-3562-46a5-87da-8561886fe273"). InnerVolumeSpecName "kube-api-access-sd5jd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.108440 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sd5jd\" (UniqueName: \"kubernetes.io/projected/d4d26cb5-3562-46a5-87da-8561886fe273-kube-api-access-sd5jd\") on node \"crc\" DevicePath \"\"" Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.108488 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4d26cb5-3562-46a5-87da-8561886fe273-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.140773 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4d26cb5-3562-46a5-87da-8561886fe273-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4d26cb5-3562-46a5-87da-8561886fe273" (UID: "d4d26cb5-3562-46a5-87da-8561886fe273"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.210278 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4d26cb5-3562-46a5-87da-8561886fe273-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.685930 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dn5ql" event={"ID":"d4d26cb5-3562-46a5-87da-8561886fe273","Type":"ContainerDied","Data":"d27e744a9f4bddc48c8263ee8da6e0966cd05333de6c62633077325fbe4cf4c6"} Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.686026 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dn5ql" Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.686044 4806 scope.go:117] "RemoveContainer" containerID="d53caf74dd1ec00020dd4bbe5ba2e34e0ada4918fbb7f286e7308f021737757e" Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.743147 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dn5ql"] Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.744462 4806 scope.go:117] "RemoveContainer" containerID="0539a73bf12da02292d191b3f92a43a52330623ca4798e15b59ea785a6392188" Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.751104 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dn5ql"] Nov 27 11:01:41 crc kubenswrapper[4806]: I1127 11:01:41.775412 4806 scope.go:117] "RemoveContainer" containerID="964af62703746475aa6be08ae0657b8f5cbc38d96ba79fa76b5248951906238c" Nov 27 11:01:42 crc kubenswrapper[4806]: I1127 11:01:42.129237 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4d26cb5-3562-46a5-87da-8561886fe273" path="/var/lib/kubelet/pods/d4d26cb5-3562-46a5-87da-8561886fe273/volumes" Nov 27 11:01:53 crc kubenswrapper[4806]: I1127 11:01:53.117211 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:01:53 crc kubenswrapper[4806]: E1127 11:01:53.118064 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:02:07 crc kubenswrapper[4806]: I1127 11:02:07.117448 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:02:07 crc kubenswrapper[4806]: E1127 11:02:07.118867 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:02:20 crc kubenswrapper[4806]: I1127 11:02:20.118610 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:02:20 crc kubenswrapper[4806]: E1127 11:02:20.119724 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:02:24 crc kubenswrapper[4806]: I1127 11:02:24.328984 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-74c97c7bb-whc9p_2f08c589-2ea3-4d79-b068-3c9784701fb0/barbican-api/0.log" Nov 27 11:02:24 crc kubenswrapper[4806]: I1127 11:02:24.763967 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-74c97c7bb-whc9p_2f08c589-2ea3-4d79-b068-3c9784701fb0/barbican-api-log/0.log" Nov 27 11:02:25 crc kubenswrapper[4806]: I1127 11:02:25.011872 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6fb499c486-z7b8m_ad03fa3a-e8e2-480d-8980-8a54cd70debb/barbican-keystone-listener/0.log" Nov 27 11:02:25 crc kubenswrapper[4806]: I1127 11:02:25.186250 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6fb499c486-z7b8m_ad03fa3a-e8e2-480d-8980-8a54cd70debb/barbican-keystone-listener-log/0.log" Nov 27 11:02:25 crc kubenswrapper[4806]: I1127 11:02:25.304160 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5bb7f69687-8mg9f_2a5920ca-ee03-41f8-9886-2c609254e7f7/barbican-worker/0.log" Nov 27 11:02:25 crc kubenswrapper[4806]: I1127 11:02:25.371275 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5bb7f69687-8mg9f_2a5920ca-ee03-41f8-9886-2c609254e7f7/barbican-worker-log/0.log" Nov 27 11:02:25 crc kubenswrapper[4806]: I1127 11:02:25.677598 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-992mx_4b39d4cf-99ff-486c-bd70-0a614a73539f/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 11:02:25 crc kubenswrapper[4806]: I1127 11:02:25.773007 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_1ebc7600-dc80-48b0-ab6b-01ba5d056c2b/ceilometer-central-agent/0.log" Nov 27 11:02:25 crc kubenswrapper[4806]: I1127 11:02:25.852147 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_1ebc7600-dc80-48b0-ab6b-01ba5d056c2b/ceilometer-notification-agent/0.log" Nov 27 11:02:25 crc kubenswrapper[4806]: I1127 11:02:25.972633 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_1ebc7600-dc80-48b0-ab6b-01ba5d056c2b/proxy-httpd/0.log" Nov 27 11:02:26 crc kubenswrapper[4806]: I1127 11:02:26.088866 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_1ebc7600-dc80-48b0-ab6b-01ba5d056c2b/sg-core/0.log" Nov 27 11:02:26 crc kubenswrapper[4806]: I1127 11:02:26.214818 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-x4dh2_1e0dc02c-8ba2-485d-97eb-664f91b36046/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 11:02:26 crc kubenswrapper[4806]: I1127 11:02:26.346569 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e91180ab-be63-4bdd-87f7-9a9975f35634/cinder-api/0.log" Nov 27 11:02:26 crc kubenswrapper[4806]: I1127 11:02:26.433715 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e91180ab-be63-4bdd-87f7-9a9975f35634/cinder-api-log/0.log" Nov 27 11:02:26 crc kubenswrapper[4806]: I1127 11:02:26.593592 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_845fe2ce-0faf-4c9a-b1f3-ce642c58a316/cinder-scheduler/0.log" Nov 27 11:02:26 crc kubenswrapper[4806]: I1127 11:02:26.782009 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_845fe2ce-0faf-4c9a-b1f3-ce642c58a316/probe/0.log" Nov 27 11:02:26 crc kubenswrapper[4806]: I1127 11:02:26.858384 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-t58pp_7339d278-f5e4-4731-a70e-bfd3735144b4/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 11:02:27 crc kubenswrapper[4806]: I1127 11:02:27.132737 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-545s4_2699b50a-1d3a-4e3b-9b01-b67f06d04f77/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 11:02:27 crc kubenswrapper[4806]: I1127 11:02:27.248405 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-667ff9c869-wcdvm_c4420782-e4e9-4241-9083-de9ace25d296/init/0.log" Nov 27 11:02:27 crc kubenswrapper[4806]: I1127 11:02:27.469298 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-667ff9c869-wcdvm_c4420782-e4e9-4241-9083-de9ace25d296/init/0.log" Nov 27 11:02:27 crc kubenswrapper[4806]: I1127 11:02:27.533721 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-667ff9c869-wcdvm_c4420782-e4e9-4241-9083-de9ace25d296/dnsmasq-dns/0.log" Nov 27 11:02:27 crc kubenswrapper[4806]: I1127 11:02:27.686213 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-gjdvq_33b9cdf1-651a-42c3-99b6-f62e1ef954bc/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 11:02:28 crc kubenswrapper[4806]: I1127 11:02:28.000456 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29404021-wxzgw_9c0737de-4b3d-42ba-8a59-068721ee9828/keystone-cron/0.log" Nov 27 11:02:28 crc kubenswrapper[4806]: I1127 11:02:28.033620 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-c9d8f7ccf-5z2jj_e28fe1bd-93d5-4138-8c4b-0a52fd6aa5f2/keystone-api/0.log" Nov 27 11:02:28 crc kubenswrapper[4806]: I1127 11:02:28.332411 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_053fa4b1-ca2c-4e00-b0ad-1d076e853dec/kube-state-metrics/0.log" Nov 27 11:02:29 crc kubenswrapper[4806]: I1127 11:02:29.270481 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c76df9f57-fglss_93660245-72f2-4792-af51-3ac612926729/neutron-api/0.log" Nov 27 11:02:29 crc kubenswrapper[4806]: I1127 11:02:29.293033 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c76df9f57-fglss_93660245-72f2-4792-af51-3ac612926729/neutron-httpd/0.log" Nov 27 11:02:29 crc kubenswrapper[4806]: I1127 11:02:29.736298 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41/nova-api-log/0.log" Nov 27 11:02:29 crc kubenswrapper[4806]: I1127 11:02:29.829527 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e99ee6cc-a2ad-47d0-b19b-b15fc1c48d41/nova-api-api/0.log" Nov 27 11:02:30 crc kubenswrapper[4806]: I1127 11:02:30.332562 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_570b138c-102e-4d40-a2df-a87abf1626b7/nova-cell0-conductor-conductor/0.log" Nov 27 11:02:30 crc kubenswrapper[4806]: I1127 11:02:30.360958 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_d93051aa-7a15-40df-92cb-7e9cef6103a8/nova-cell1-conductor-conductor/0.log" Nov 27 11:02:30 crc kubenswrapper[4806]: I1127 11:02:30.694143 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_aa061b40-cf89-478c-a82e-957785f57c69/nova-cell1-novncproxy-novncproxy/0.log" Nov 27 11:02:30 crc kubenswrapper[4806]: I1127 11:02:30.791375 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_279060f3-fe83-4add-9a41-d5bf7e4f4e4a/nova-metadata-log/0.log" Nov 27 11:02:31 crc kubenswrapper[4806]: I1127 11:02:31.144505 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_69605757-d5b0-4169-9c6e-15f51f282bb3/nova-scheduler-scheduler/0.log" Nov 27 11:02:31 crc kubenswrapper[4806]: I1127 11:02:31.439810 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_46cb1b0e-a6ab-467a-9f49-f8938c164253/mysql-bootstrap/0.log" Nov 27 11:02:31 crc kubenswrapper[4806]: I1127 11:02:31.505348 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_279060f3-fe83-4add-9a41-d5bf7e4f4e4a/nova-metadata-metadata/0.log" Nov 27 11:02:31 crc kubenswrapper[4806]: I1127 11:02:31.760999 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_46cb1b0e-a6ab-467a-9f49-f8938c164253/galera/0.log" Nov 27 11:02:31 crc kubenswrapper[4806]: I1127 11:02:31.791802 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b30ea2f1-af76-4a43-842c-577ff4f35a20/mysql-bootstrap/0.log" Nov 27 11:02:31 crc kubenswrapper[4806]: I1127 11:02:31.838167 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_46cb1b0e-a6ab-467a-9f49-f8938c164253/mysql-bootstrap/0.log" Nov 27 11:02:32 crc kubenswrapper[4806]: I1127 11:02:32.149054 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b30ea2f1-af76-4a43-842c-577ff4f35a20/mysql-bootstrap/0.log" Nov 27 11:02:32 crc kubenswrapper[4806]: I1127 11:02:32.192334 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b30ea2f1-af76-4a43-842c-577ff4f35a20/galera/0.log" Nov 27 11:02:32 crc kubenswrapper[4806]: I1127 11:02:32.264003 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_01bb064b-2da1-4695-8e6e-8c635ff6dbaa/openstackclient/0.log" Nov 27 11:02:32 crc kubenswrapper[4806]: I1127 11:02:32.606667 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-l82sn_2cb892e8-358b-477b-8741-4211a599bf6c/ovn-controller/0.log" Nov 27 11:02:32 crc kubenswrapper[4806]: I1127 11:02:32.839682 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-xg9f9_a81c0343-88d6-42cf-bc8f-5a3b8722a137/openstack-network-exporter/0.log" Nov 27 11:02:33 crc kubenswrapper[4806]: I1127 11:02:33.121095 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:02:33 crc kubenswrapper[4806]: E1127 11:02:33.121670 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:02:33 crc kubenswrapper[4806]: I1127 11:02:33.525607 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-f769m_d814f280-e3ac-426c-bafa-4a4ed3f51320/ovsdb-server-init/0.log" Nov 27 11:02:33 crc kubenswrapper[4806]: I1127 11:02:33.683454 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-f769m_d814f280-e3ac-426c-bafa-4a4ed3f51320/ovs-vswitchd/0.log" Nov 27 11:02:33 crc kubenswrapper[4806]: I1127 11:02:33.841024 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-f769m_d814f280-e3ac-426c-bafa-4a4ed3f51320/ovsdb-server-init/0.log" Nov 27 11:02:33 crc kubenswrapper[4806]: I1127 11:02:33.871499 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-f769m_d814f280-e3ac-426c-bafa-4a4ed3f51320/ovsdb-server/0.log" Nov 27 11:02:34 crc kubenswrapper[4806]: I1127 11:02:34.048766 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5a10ac5e-a435-45ef-b143-f6578010557f/openstack-network-exporter/0.log" Nov 27 11:02:34 crc kubenswrapper[4806]: I1127 11:02:34.128182 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5a10ac5e-a435-45ef-b143-f6578010557f/ovn-northd/0.log" Nov 27 11:02:34 crc kubenswrapper[4806]: I1127 11:02:34.256515 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_08068118-444f-4ea3-9c3c-7f0768a257b2/openstack-network-exporter/0.log" Nov 27 11:02:34 crc kubenswrapper[4806]: I1127 11:02:34.455421 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_08068118-444f-4ea3-9c3c-7f0768a257b2/ovsdbserver-nb/0.log" Nov 27 11:02:34 crc kubenswrapper[4806]: I1127 11:02:34.566557 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d8134ca5-1cd3-4793-875e-8ea7bde9c407/openstack-network-exporter/0.log" Nov 27 11:02:34 crc kubenswrapper[4806]: I1127 11:02:34.691814 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d8134ca5-1cd3-4793-875e-8ea7bde9c407/ovsdbserver-sb/0.log" Nov 27 11:02:35 crc kubenswrapper[4806]: I1127 11:02:35.022676 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5689867474-qpmnv_9e42e515-17c4-4306-922d-665279ec1c3f/placement-api/0.log" Nov 27 11:02:35 crc kubenswrapper[4806]: I1127 11:02:35.077343 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5689867474-qpmnv_9e42e515-17c4-4306-922d-665279ec1c3f/placement-log/0.log" Nov 27 11:02:35 crc kubenswrapper[4806]: I1127 11:02:35.226605 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3cdb0a58-54e1-4d11-8240-721304f548dd/setup-container/0.log" Nov 27 11:02:35 crc kubenswrapper[4806]: I1127 11:02:35.614038 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3cdb0a58-54e1-4d11-8240-721304f548dd/setup-container/0.log" Nov 27 11:02:35 crc kubenswrapper[4806]: I1127 11:02:35.674145 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5052338d-2912-432a-a022-aa169c1d0930/setup-container/0.log" Nov 27 11:02:35 crc kubenswrapper[4806]: I1127 11:02:35.687245 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3cdb0a58-54e1-4d11-8240-721304f548dd/rabbitmq/0.log" Nov 27 11:02:36 crc kubenswrapper[4806]: I1127 11:02:36.023578 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5052338d-2912-432a-a022-aa169c1d0930/setup-container/0.log" Nov 27 11:02:36 crc kubenswrapper[4806]: I1127 11:02:36.064682 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-z75td_078a0300-a31f-4b03-ba13-66692b945cb2/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 11:02:36 crc kubenswrapper[4806]: I1127 11:02:36.122366 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5052338d-2912-432a-a022-aa169c1d0930/rabbitmq/0.log" Nov 27 11:02:36 crc kubenswrapper[4806]: I1127 11:02:36.262823 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_d9a162dd-03ba-4444-b03c-bf34c535713a/memcached/0.log" Nov 27 11:02:36 crc kubenswrapper[4806]: I1127 11:02:36.359964 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-9862t_8969ea2b-0bca-42aa-977a-4897c1ce29fb/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 11:02:36 crc kubenswrapper[4806]: I1127 11:02:36.592020 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-qvprb_c358c0e7-2566-4f18-8a4a-2af284af91f6/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 11:02:36 crc kubenswrapper[4806]: I1127 11:02:36.981415 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-djqnt_029a4e3d-528c-47a0-b34d-42f296055ed0/ssh-known-hosts-edpm-deployment/0.log" Nov 27 11:02:37 crc kubenswrapper[4806]: I1127 11:02:37.064638 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-gw5sb_7197a402-7e1b-450a-ba53-448bf843e6f6/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 27 11:02:47 crc kubenswrapper[4806]: I1127 11:02:47.117370 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:02:47 crc kubenswrapper[4806]: E1127 11:02:47.118309 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:03:01 crc kubenswrapper[4806]: I1127 11:03:01.124914 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:03:01 crc kubenswrapper[4806]: E1127 11:03:01.126652 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:03:02 crc kubenswrapper[4806]: I1127 11:03:02.898439 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-kn7j2_96cf5347-61b7-4ecb-8093-59b25482cd54/kube-rbac-proxy/0.log" Nov 27 11:03:03 crc kubenswrapper[4806]: I1127 11:03:03.108680 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-kn7j2_96cf5347-61b7-4ecb-8093-59b25482cd54/manager/0.log" Nov 27 11:03:03 crc kubenswrapper[4806]: I1127 11:03:03.351214 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-lcsw6_33c0104b-4a8a-46fa-bd83-b42f86f96789/kube-rbac-proxy/0.log" Nov 27 11:03:03 crc kubenswrapper[4806]: I1127 11:03:03.431866 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-lcsw6_33c0104b-4a8a-46fa-bd83-b42f86f96789/manager/0.log" Nov 27 11:03:03 crc kubenswrapper[4806]: I1127 11:03:03.472026 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq_99394f5f-eef5-416a-ae48-376b3e6f6122/util/0.log" Nov 27 11:03:03 crc kubenswrapper[4806]: I1127 11:03:03.689328 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq_99394f5f-eef5-416a-ae48-376b3e6f6122/pull/0.log" Nov 27 11:03:03 crc kubenswrapper[4806]: I1127 11:03:03.717957 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq_99394f5f-eef5-416a-ae48-376b3e6f6122/util/0.log" Nov 27 11:03:03 crc kubenswrapper[4806]: I1127 11:03:03.729974 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq_99394f5f-eef5-416a-ae48-376b3e6f6122/pull/0.log" Nov 27 11:03:03 crc kubenswrapper[4806]: I1127 11:03:03.945306 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq_99394f5f-eef5-416a-ae48-376b3e6f6122/util/0.log" Nov 27 11:03:03 crc kubenswrapper[4806]: I1127 11:03:03.963943 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq_99394f5f-eef5-416a-ae48-376b3e6f6122/pull/0.log" Nov 27 11:03:04 crc kubenswrapper[4806]: I1127 11:03:04.057757 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d8d7dd9ae4e96a2f3fcd790ea7044d20eb1ef07c8d10a3c9798653913bmgdpq_99394f5f-eef5-416a-ae48-376b3e6f6122/extract/0.log" Nov 27 11:03:04 crc kubenswrapper[4806]: I1127 11:03:04.214258 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-ct9ps_93cca2e7-6982-42d3-8848-10e4f7a70509/kube-rbac-proxy/0.log" Nov 27 11:03:04 crc kubenswrapper[4806]: I1127 11:03:04.322899 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-9jh67_b6163410-3453-43c3-8cd7-5a354e447025/kube-rbac-proxy/0.log" Nov 27 11:03:04 crc kubenswrapper[4806]: I1127 11:03:04.332296 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-ct9ps_93cca2e7-6982-42d3-8848-10e4f7a70509/manager/0.log" Nov 27 11:03:04 crc kubenswrapper[4806]: I1127 11:03:04.545412 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-9jh67_b6163410-3453-43c3-8cd7-5a354e447025/manager/0.log" Nov 27 11:03:04 crc kubenswrapper[4806]: I1127 11:03:04.638334 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-czdn8_94f64657-d78c-4f52-8aa1-b23b61d5af68/manager/0.log" Nov 27 11:03:04 crc kubenswrapper[4806]: I1127 11:03:04.643120 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-czdn8_94f64657-d78c-4f52-8aa1-b23b61d5af68/kube-rbac-proxy/0.log" Nov 27 11:03:04 crc kubenswrapper[4806]: I1127 11:03:04.871702 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-v8r26_5b227bdd-2c69-4053-a116-8085cd157a93/kube-rbac-proxy/0.log" Nov 27 11:03:04 crc kubenswrapper[4806]: I1127 11:03:04.926013 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-v8r26_5b227bdd-2c69-4053-a116-8085cd157a93/manager/0.log" Nov 27 11:03:05 crc kubenswrapper[4806]: I1127 11:03:05.187113 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-nkrwj_609248ce-9b50-48c5-90bd-aa3d2c8e2401/kube-rbac-proxy/0.log" Nov 27 11:03:05 crc kubenswrapper[4806]: I1127 11:03:05.192961 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-55nb9_60db26f8-5bc6-477c-bf21-6842e1b3a9db/kube-rbac-proxy/0.log" Nov 27 11:03:05 crc kubenswrapper[4806]: I1127 11:03:05.216727 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-55nb9_60db26f8-5bc6-477c-bf21-6842e1b3a9db/manager/0.log" Nov 27 11:03:05 crc kubenswrapper[4806]: I1127 11:03:05.420451 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-m6mcb_856215a1-cdc4-49fb-a522-0c2ca14cd757/kube-rbac-proxy/0.log" Nov 27 11:03:05 crc kubenswrapper[4806]: I1127 11:03:05.421811 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-nkrwj_609248ce-9b50-48c5-90bd-aa3d2c8e2401/manager/0.log" Nov 27 11:03:05 crc kubenswrapper[4806]: I1127 11:03:05.556480 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-m6mcb_856215a1-cdc4-49fb-a522-0c2ca14cd757/manager/0.log" Nov 27 11:03:05 crc kubenswrapper[4806]: I1127 11:03:05.667667 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-4vz5m_4e3d3cf9-5e6d-4963-a2b2-731f4fb5ecb2/kube-rbac-proxy/0.log" Nov 27 11:03:05 crc kubenswrapper[4806]: I1127 11:03:05.706332 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-4vz5m_4e3d3cf9-5e6d-4963-a2b2-731f4fb5ecb2/manager/0.log" Nov 27 11:03:05 crc kubenswrapper[4806]: I1127 11:03:05.926991 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-j2x9t_671409ea-bf88-4550-a3f7-1fb63a3e89ed/kube-rbac-proxy/0.log" Nov 27 11:03:05 crc kubenswrapper[4806]: I1127 11:03:05.986979 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-j2x9t_671409ea-bf88-4550-a3f7-1fb63a3e89ed/manager/0.log" Nov 27 11:03:06 crc kubenswrapper[4806]: I1127 11:03:06.240331 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-tdgzm_4740d40d-863a-40a0-a46f-b515f10e4733/manager/0.log" Nov 27 11:03:06 crc kubenswrapper[4806]: I1127 11:03:06.257726 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-jtsbv_35e9f707-4424-43fc-8d97-bb804524a180/kube-rbac-proxy/0.log" Nov 27 11:03:06 crc kubenswrapper[4806]: I1127 11:03:06.347474 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-tdgzm_4740d40d-863a-40a0-a46f-b515f10e4733/kube-rbac-proxy/0.log" Nov 27 11:03:06 crc kubenswrapper[4806]: I1127 11:03:06.513013 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-jtsbv_35e9f707-4424-43fc-8d97-bb804524a180/manager/0.log" Nov 27 11:03:06 crc kubenswrapper[4806]: I1127 11:03:06.571082 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-hpmjm_e5d637d7-a046-4026-a157-abcb8d99ba3b/kube-rbac-proxy/0.log" Nov 27 11:03:06 crc kubenswrapper[4806]: I1127 11:03:06.627341 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-hpmjm_e5d637d7-a046-4026-a157-abcb8d99ba3b/manager/0.log" Nov 27 11:03:06 crc kubenswrapper[4806]: I1127 11:03:06.832382 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh_55fdd641-f8c1-417f-8abb-8adfc95d7008/kube-rbac-proxy/0.log" Nov 27 11:03:06 crc kubenswrapper[4806]: I1127 11:03:06.887458 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6b88gvh_55fdd641-f8c1-417f-8abb-8adfc95d7008/manager/0.log" Nov 27 11:03:07 crc kubenswrapper[4806]: I1127 11:03:07.252672 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6784745db8-rng9q_33c3ee1d-c0cc-4c75-b6e9-0059dd8042c7/operator/0.log" Nov 27 11:03:07 crc kubenswrapper[4806]: I1127 11:03:07.379738 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-qjct2_7a4448af-a960-4299-8ed2-71b93c8b7e46/registry-server/0.log" Nov 27 11:03:07 crc kubenswrapper[4806]: I1127 11:03:07.397623 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-82qfv_159a9bdd-85be-42dc-8d60-19c3eb1a90e4/kube-rbac-proxy/0.log" Nov 27 11:03:07 crc kubenswrapper[4806]: I1127 11:03:07.632025 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-82qfv_159a9bdd-85be-42dc-8d60-19c3eb1a90e4/manager/0.log" Nov 27 11:03:07 crc kubenswrapper[4806]: I1127 11:03:07.669850 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-hrbw7_92d56580-f256-46f7-81fa-e370048b4d0f/kube-rbac-proxy/0.log" Nov 27 11:03:07 crc kubenswrapper[4806]: I1127 11:03:07.827767 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-hrbw7_92d56580-f256-46f7-81fa-e370048b4d0f/manager/0.log" Nov 27 11:03:07 crc kubenswrapper[4806]: I1127 11:03:07.917490 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7fbdc78c67-7th5m_a11f64de-02ce-4b43-9ed8-54d0c844e48e/manager/0.log" Nov 27 11:03:07 crc kubenswrapper[4806]: I1127 11:03:07.972675 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-bz87k_3104471d-6a15-4454-863e-b5754f748043/operator/0.log" Nov 27 11:03:08 crc kubenswrapper[4806]: I1127 11:03:08.164159 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-xs5lm_f77adb32-dfac-48de-b4a8-c9263cf5729d/kube-rbac-proxy/0.log" Nov 27 11:03:08 crc kubenswrapper[4806]: I1127 11:03:08.172688 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-xs5lm_f77adb32-dfac-48de-b4a8-c9263cf5729d/manager/0.log" Nov 27 11:03:08 crc kubenswrapper[4806]: I1127 11:03:08.257621 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-fjhns_d1ceaced-2100-4952-9bc8-cbd5c82886c5/kube-rbac-proxy/0.log" Nov 27 11:03:08 crc kubenswrapper[4806]: I1127 11:03:08.373133 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-fjhns_d1ceaced-2100-4952-9bc8-cbd5c82886c5/manager/0.log" Nov 27 11:03:08 crc kubenswrapper[4806]: I1127 11:03:08.482039 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-6fkj4_08fa2215-67c9-4110-b54d-7a18e1641d88/manager/0.log" Nov 27 11:03:08 crc kubenswrapper[4806]: I1127 11:03:08.503272 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-6fkj4_08fa2215-67c9-4110-b54d-7a18e1641d88/kube-rbac-proxy/0.log" Nov 27 11:03:08 crc kubenswrapper[4806]: I1127 11:03:08.655306 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-v7tst_b02aea7d-ec4b-425f-8d13-6205c91ee589/kube-rbac-proxy/0.log" Nov 27 11:03:08 crc kubenswrapper[4806]: I1127 11:03:08.696160 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-v7tst_b02aea7d-ec4b-425f-8d13-6205c91ee589/manager/0.log" Nov 27 11:03:12 crc kubenswrapper[4806]: I1127 11:03:12.117780 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:03:12 crc kubenswrapper[4806]: E1127 11:03:12.118562 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:03:27 crc kubenswrapper[4806]: I1127 11:03:27.117402 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:03:28 crc kubenswrapper[4806]: I1127 11:03:28.258027 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"3a74cf9f736bfbb3f039de322c4fc66a3aea02cb5343e605f8777b301781b8fe"} Nov 27 11:03:30 crc kubenswrapper[4806]: I1127 11:03:30.101952 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5qtlg_adc0ccaf-fcb4-4aa9-8b3a-9b1e08c3b1ce/control-plane-machine-set-operator/0.log" Nov 27 11:03:30 crc kubenswrapper[4806]: I1127 11:03:30.375153 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zd4hv_325e82c6-eacf-4180-bc5f-9334536ebe16/kube-rbac-proxy/0.log" Nov 27 11:03:30 crc kubenswrapper[4806]: I1127 11:03:30.381597 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zd4hv_325e82c6-eacf-4180-bc5f-9334536ebe16/machine-api-operator/0.log" Nov 27 11:03:43 crc kubenswrapper[4806]: I1127 11:03:43.593115 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-pvxdj_2e8a774f-7876-4f56-94e6-c559397129c1/cert-manager-controller/0.log" Nov 27 11:03:43 crc kubenswrapper[4806]: I1127 11:03:43.774811 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-6dclx_42836747-f4cd-43a2-a690-b17989a8d2c1/cert-manager-cainjector/0.log" Nov 27 11:03:43 crc kubenswrapper[4806]: I1127 11:03:43.838038 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-ptpz9_9f1cba27-3a94-461f-bee4-1dde7d31f9a6/cert-manager-webhook/0.log" Nov 27 11:03:58 crc kubenswrapper[4806]: I1127 11:03:58.548768 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-bncdw_733b60f5-7297-420d-b952-896a3156f5f7/nmstate-console-plugin/0.log" Nov 27 11:03:58 crc kubenswrapper[4806]: I1127 11:03:58.837509 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-979gk_385b9d19-bcb8-40b8-ab88-7711ed6070e8/nmstate-handler/0.log" Nov 27 11:03:58 crc kubenswrapper[4806]: I1127 11:03:58.936898 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-np48l_412527c7-8920-4491-9035-2af48be2224b/kube-rbac-proxy/0.log" Nov 27 11:03:58 crc kubenswrapper[4806]: I1127 11:03:58.976527 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-np48l_412527c7-8920-4491-9035-2af48be2224b/nmstate-metrics/0.log" Nov 27 11:03:59 crc kubenswrapper[4806]: I1127 11:03:59.274190 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-mtxjk_53a13ef4-b142-48f4-b6ae-81d8fb9a7203/nmstate-operator/0.log" Nov 27 11:03:59 crc kubenswrapper[4806]: I1127 11:03:59.284125 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-lv2hq_29deb9cb-7348-4dba-b3d3-cfbfdb88a79f/nmstate-webhook/0.log" Nov 27 11:04:16 crc kubenswrapper[4806]: I1127 11:04:16.288406 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-7c7qd_d73ced61-9b8c-475c-b971-362279798b8e/kube-rbac-proxy/0.log" Nov 27 11:04:16 crc kubenswrapper[4806]: I1127 11:04:16.454700 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-7c7qd_d73ced61-9b8c-475c-b971-362279798b8e/controller/0.log" Nov 27 11:04:16 crc kubenswrapper[4806]: I1127 11:04:16.607001 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/cp-frr-files/0.log" Nov 27 11:04:16 crc kubenswrapper[4806]: I1127 11:04:16.770617 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/cp-frr-files/0.log" Nov 27 11:04:16 crc kubenswrapper[4806]: I1127 11:04:16.861428 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/cp-metrics/0.log" Nov 27 11:04:16 crc kubenswrapper[4806]: I1127 11:04:16.879170 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/cp-reloader/0.log" Nov 27 11:04:16 crc kubenswrapper[4806]: I1127 11:04:16.936335 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/cp-reloader/0.log" Nov 27 11:04:17 crc kubenswrapper[4806]: I1127 11:04:17.163986 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/cp-metrics/0.log" Nov 27 11:04:17 crc kubenswrapper[4806]: I1127 11:04:17.167099 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/cp-frr-files/0.log" Nov 27 11:04:17 crc kubenswrapper[4806]: I1127 11:04:17.168212 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/cp-metrics/0.log" Nov 27 11:04:17 crc kubenswrapper[4806]: I1127 11:04:17.184378 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/cp-reloader/0.log" Nov 27 11:04:17 crc kubenswrapper[4806]: I1127 11:04:17.490246 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/cp-metrics/0.log" Nov 27 11:04:17 crc kubenswrapper[4806]: I1127 11:04:17.492417 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/controller/0.log" Nov 27 11:04:17 crc kubenswrapper[4806]: I1127 11:04:17.523838 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/cp-frr-files/0.log" Nov 27 11:04:17 crc kubenswrapper[4806]: I1127 11:04:17.563523 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/cp-reloader/0.log" Nov 27 11:04:17 crc kubenswrapper[4806]: I1127 11:04:17.806329 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/kube-rbac-proxy/0.log" Nov 27 11:04:17 crc kubenswrapper[4806]: I1127 11:04:17.860487 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/frr-metrics/0.log" Nov 27 11:04:17 crc kubenswrapper[4806]: I1127 11:04:17.860842 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/kube-rbac-proxy-frr/0.log" Nov 27 11:04:18 crc kubenswrapper[4806]: I1127 11:04:18.165008 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/reloader/0.log" Nov 27 11:04:18 crc kubenswrapper[4806]: I1127 11:04:18.237374 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-l4ld2_ed9cee58-4b54-4ef2-a86c-99c8f0e83d8b/frr-k8s-webhook-server/0.log" Nov 27 11:04:18 crc kubenswrapper[4806]: I1127 11:04:18.464631 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-766c64499f-pl7sk_d62398ac-46b4-4e4f-b3f9-9ae58756b4b3/manager/0.log" Nov 27 11:04:18 crc kubenswrapper[4806]: I1127 11:04:18.789104 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd27h_5e7e1847-fefb-4289-8dbb-c04dec670021/frr/0.log" Nov 27 11:04:18 crc kubenswrapper[4806]: I1127 11:04:18.792095 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7cbfdf6f8b-zwpmf_020fead9-ad47-4930-8797-88254cbc0245/webhook-server/0.log" Nov 27 11:04:18 crc kubenswrapper[4806]: I1127 11:04:18.923708 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bvd28_ec562de0-63d3-4613-96f6-b44435bd7dab/kube-rbac-proxy/0.log" Nov 27 11:04:19 crc kubenswrapper[4806]: I1127 11:04:19.327433 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bvd28_ec562de0-63d3-4613-96f6-b44435bd7dab/speaker/0.log" Nov 27 11:04:34 crc kubenswrapper[4806]: I1127 11:04:34.026388 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2_4dd8e785-4b1d-406d-bb97-391b74415228/util/0.log" Nov 27 11:04:34 crc kubenswrapper[4806]: I1127 11:04:34.166524 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2_4dd8e785-4b1d-406d-bb97-391b74415228/util/0.log" Nov 27 11:04:34 crc kubenswrapper[4806]: I1127 11:04:34.205259 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2_4dd8e785-4b1d-406d-bb97-391b74415228/pull/0.log" Nov 27 11:04:34 crc kubenswrapper[4806]: I1127 11:04:34.283727 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2_4dd8e785-4b1d-406d-bb97-391b74415228/pull/0.log" Nov 27 11:04:34 crc kubenswrapper[4806]: I1127 11:04:34.566530 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2_4dd8e785-4b1d-406d-bb97-391b74415228/util/0.log" Nov 27 11:04:34 crc kubenswrapper[4806]: I1127 11:04:34.696519 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2_4dd8e785-4b1d-406d-bb97-391b74415228/pull/0.log" Nov 27 11:04:34 crc kubenswrapper[4806]: I1127 11:04:34.735563 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fbksm2_4dd8e785-4b1d-406d-bb97-391b74415228/extract/0.log" Nov 27 11:04:35 crc kubenswrapper[4806]: I1127 11:04:35.239179 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd_2afa20bc-c32f-4187-90b9-f41cf646cc8a/util/0.log" Nov 27 11:04:35 crc kubenswrapper[4806]: I1127 11:04:35.528292 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd_2afa20bc-c32f-4187-90b9-f41cf646cc8a/pull/0.log" Nov 27 11:04:35 crc kubenswrapper[4806]: I1127 11:04:35.675024 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd_2afa20bc-c32f-4187-90b9-f41cf646cc8a/pull/0.log" Nov 27 11:04:35 crc kubenswrapper[4806]: I1127 11:04:35.682448 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd_2afa20bc-c32f-4187-90b9-f41cf646cc8a/util/0.log" Nov 27 11:04:35 crc kubenswrapper[4806]: I1127 11:04:35.914584 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd_2afa20bc-c32f-4187-90b9-f41cf646cc8a/extract/0.log" Nov 27 11:04:35 crc kubenswrapper[4806]: I1127 11:04:35.955417 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd_2afa20bc-c32f-4187-90b9-f41cf646cc8a/util/0.log" Nov 27 11:04:36 crc kubenswrapper[4806]: I1127 11:04:36.016404 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83q4lxd_2afa20bc-c32f-4187-90b9-f41cf646cc8a/pull/0.log" Nov 27 11:04:36 crc kubenswrapper[4806]: I1127 11:04:36.192153 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2d5nd_525641e9-e2fb-4862-8fe5-ada748dd490e/extract-utilities/0.log" Nov 27 11:04:36 crc kubenswrapper[4806]: I1127 11:04:36.469908 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2d5nd_525641e9-e2fb-4862-8fe5-ada748dd490e/extract-utilities/0.log" Nov 27 11:04:36 crc kubenswrapper[4806]: I1127 11:04:36.536479 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2d5nd_525641e9-e2fb-4862-8fe5-ada748dd490e/extract-content/0.log" Nov 27 11:04:36 crc kubenswrapper[4806]: I1127 11:04:36.560512 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2d5nd_525641e9-e2fb-4862-8fe5-ada748dd490e/extract-content/0.log" Nov 27 11:04:36 crc kubenswrapper[4806]: I1127 11:04:36.803594 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2d5nd_525641e9-e2fb-4862-8fe5-ada748dd490e/extract-utilities/0.log" Nov 27 11:04:36 crc kubenswrapper[4806]: I1127 11:04:36.810828 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2d5nd_525641e9-e2fb-4862-8fe5-ada748dd490e/extract-content/0.log" Nov 27 11:04:37 crc kubenswrapper[4806]: I1127 11:04:37.051803 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pc2bp_a19e7111-3ef1-45a9-ae31-0d53a2a6615f/extract-utilities/0.log" Nov 27 11:04:37 crc kubenswrapper[4806]: I1127 11:04:37.158833 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2d5nd_525641e9-e2fb-4862-8fe5-ada748dd490e/registry-server/0.log" Nov 27 11:04:37 crc kubenswrapper[4806]: I1127 11:04:37.472119 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pc2bp_a19e7111-3ef1-45a9-ae31-0d53a2a6615f/extract-content/0.log" Nov 27 11:04:37 crc kubenswrapper[4806]: I1127 11:04:37.473926 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pc2bp_a19e7111-3ef1-45a9-ae31-0d53a2a6615f/extract-content/0.log" Nov 27 11:04:37 crc kubenswrapper[4806]: I1127 11:04:37.534159 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pc2bp_a19e7111-3ef1-45a9-ae31-0d53a2a6615f/extract-utilities/0.log" Nov 27 11:04:37 crc kubenswrapper[4806]: I1127 11:04:37.772362 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pc2bp_a19e7111-3ef1-45a9-ae31-0d53a2a6615f/extract-content/0.log" Nov 27 11:04:37 crc kubenswrapper[4806]: I1127 11:04:37.817521 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pc2bp_a19e7111-3ef1-45a9-ae31-0d53a2a6615f/extract-utilities/0.log" Nov 27 11:04:38 crc kubenswrapper[4806]: I1127 11:04:38.161096 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xcj48_4e80f08f-5327-4a27-a182-f26bf02f2f17/marketplace-operator/0.log" Nov 27 11:04:38 crc kubenswrapper[4806]: I1127 11:04:38.176093 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pc2bp_a19e7111-3ef1-45a9-ae31-0d53a2a6615f/registry-server/0.log" Nov 27 11:04:38 crc kubenswrapper[4806]: I1127 11:04:38.656286 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j54w9_5e9f754d-2097-4526-9123-757c1af2a591/extract-utilities/0.log" Nov 27 11:04:38 crc kubenswrapper[4806]: I1127 11:04:38.983400 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j54w9_5e9f754d-2097-4526-9123-757c1af2a591/extract-utilities/0.log" Nov 27 11:04:39 crc kubenswrapper[4806]: I1127 11:04:39.005849 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j54w9_5e9f754d-2097-4526-9123-757c1af2a591/extract-content/0.log" Nov 27 11:04:39 crc kubenswrapper[4806]: I1127 11:04:39.016351 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j54w9_5e9f754d-2097-4526-9123-757c1af2a591/extract-content/0.log" Nov 27 11:04:39 crc kubenswrapper[4806]: I1127 11:04:39.283945 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j54w9_5e9f754d-2097-4526-9123-757c1af2a591/extract-utilities/0.log" Nov 27 11:04:39 crc kubenswrapper[4806]: I1127 11:04:39.315173 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j54w9_5e9f754d-2097-4526-9123-757c1af2a591/extract-content/0.log" Nov 27 11:04:39 crc kubenswrapper[4806]: I1127 11:04:39.346973 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-69nx8_eff0fde1-b975-41bc-95a4-080177acf5ef/extract-utilities/0.log" Nov 27 11:04:39 crc kubenswrapper[4806]: I1127 11:04:39.462837 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-j54w9_5e9f754d-2097-4526-9123-757c1af2a591/registry-server/0.log" Nov 27 11:04:39 crc kubenswrapper[4806]: I1127 11:04:39.665957 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-69nx8_eff0fde1-b975-41bc-95a4-080177acf5ef/extract-content/0.log" Nov 27 11:04:39 crc kubenswrapper[4806]: I1127 11:04:39.666634 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-69nx8_eff0fde1-b975-41bc-95a4-080177acf5ef/extract-utilities/0.log" Nov 27 11:04:39 crc kubenswrapper[4806]: I1127 11:04:39.666741 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-69nx8_eff0fde1-b975-41bc-95a4-080177acf5ef/extract-content/0.log" Nov 27 11:04:40 crc kubenswrapper[4806]: I1127 11:04:40.054203 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-69nx8_eff0fde1-b975-41bc-95a4-080177acf5ef/extract-utilities/0.log" Nov 27 11:04:40 crc kubenswrapper[4806]: I1127 11:04:40.089251 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-69nx8_eff0fde1-b975-41bc-95a4-080177acf5ef/extract-content/0.log" Nov 27 11:04:40 crc kubenswrapper[4806]: I1127 11:04:40.264568 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-69nx8_eff0fde1-b975-41bc-95a4-080177acf5ef/registry-server/0.log" Nov 27 11:05:44 crc kubenswrapper[4806]: I1127 11:05:44.450428 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:05:44 crc kubenswrapper[4806]: I1127 11:05:44.451075 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:06:14 crc kubenswrapper[4806]: I1127 11:06:14.451060 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:06:14 crc kubenswrapper[4806]: I1127 11:06:14.451973 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.576052 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5ms5g"] Nov 27 11:06:19 crc kubenswrapper[4806]: E1127 11:06:19.580931 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66620702-1b55-49ef-babc-4d43a8767274" containerName="container-00" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.581100 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="66620702-1b55-49ef-babc-4d43a8767274" containerName="container-00" Nov 27 11:06:19 crc kubenswrapper[4806]: E1127 11:06:19.581253 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4d26cb5-3562-46a5-87da-8561886fe273" containerName="extract-utilities" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.581373 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4d26cb5-3562-46a5-87da-8561886fe273" containerName="extract-utilities" Nov 27 11:06:19 crc kubenswrapper[4806]: E1127 11:06:19.581501 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4d26cb5-3562-46a5-87da-8561886fe273" containerName="registry-server" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.581610 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4d26cb5-3562-46a5-87da-8561886fe273" containerName="registry-server" Nov 27 11:06:19 crc kubenswrapper[4806]: E1127 11:06:19.581742 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4d26cb5-3562-46a5-87da-8561886fe273" containerName="extract-content" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.582103 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4d26cb5-3562-46a5-87da-8561886fe273" containerName="extract-content" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.582590 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="66620702-1b55-49ef-babc-4d43a8767274" containerName="container-00" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.582753 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4d26cb5-3562-46a5-87da-8561886fe273" containerName="registry-server" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.586053 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.601465 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ms5g"] Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.660848 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e4632ba-3288-40e2-adf8-671fe25a869b-utilities\") pod \"redhat-marketplace-5ms5g\" (UID: \"5e4632ba-3288-40e2-adf8-671fe25a869b\") " pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.661186 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e4632ba-3288-40e2-adf8-671fe25a869b-catalog-content\") pod \"redhat-marketplace-5ms5g\" (UID: \"5e4632ba-3288-40e2-adf8-671fe25a869b\") " pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.661325 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcpfc\" (UniqueName: \"kubernetes.io/projected/5e4632ba-3288-40e2-adf8-671fe25a869b-kube-api-access-zcpfc\") pod \"redhat-marketplace-5ms5g\" (UID: \"5e4632ba-3288-40e2-adf8-671fe25a869b\") " pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.762790 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e4632ba-3288-40e2-adf8-671fe25a869b-utilities\") pod \"redhat-marketplace-5ms5g\" (UID: \"5e4632ba-3288-40e2-adf8-671fe25a869b\") " pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.762879 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e4632ba-3288-40e2-adf8-671fe25a869b-catalog-content\") pod \"redhat-marketplace-5ms5g\" (UID: \"5e4632ba-3288-40e2-adf8-671fe25a869b\") " pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.762923 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcpfc\" (UniqueName: \"kubernetes.io/projected/5e4632ba-3288-40e2-adf8-671fe25a869b-kube-api-access-zcpfc\") pod \"redhat-marketplace-5ms5g\" (UID: \"5e4632ba-3288-40e2-adf8-671fe25a869b\") " pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.763470 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e4632ba-3288-40e2-adf8-671fe25a869b-utilities\") pod \"redhat-marketplace-5ms5g\" (UID: \"5e4632ba-3288-40e2-adf8-671fe25a869b\") " pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.763569 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e4632ba-3288-40e2-adf8-671fe25a869b-catalog-content\") pod \"redhat-marketplace-5ms5g\" (UID: \"5e4632ba-3288-40e2-adf8-671fe25a869b\") " pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.792779 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcpfc\" (UniqueName: \"kubernetes.io/projected/5e4632ba-3288-40e2-adf8-671fe25a869b-kube-api-access-zcpfc\") pod \"redhat-marketplace-5ms5g\" (UID: \"5e4632ba-3288-40e2-adf8-671fe25a869b\") " pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:19 crc kubenswrapper[4806]: I1127 11:06:19.953749 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:20 crc kubenswrapper[4806]: I1127 11:06:20.377495 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ms5g"] Nov 27 11:06:20 crc kubenswrapper[4806]: E1127 11:06:20.975774 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e4632ba_3288_40e2_adf8_671fe25a869b.slice/crio-e6591f36c50290a92d87c5edd46e3c0eff59a4fbe62b90cb88b3aae391353633.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e4632ba_3288_40e2_adf8_671fe25a869b.slice/crio-conmon-e6591f36c50290a92d87c5edd46e3c0eff59a4fbe62b90cb88b3aae391353633.scope\": RecentStats: unable to find data in memory cache]" Nov 27 11:06:21 crc kubenswrapper[4806]: I1127 11:06:21.035308 4806 generic.go:334] "Generic (PLEG): container finished" podID="5e4632ba-3288-40e2-adf8-671fe25a869b" containerID="e6591f36c50290a92d87c5edd46e3c0eff59a4fbe62b90cb88b3aae391353633" exitCode=0 Nov 27 11:06:21 crc kubenswrapper[4806]: I1127 11:06:21.035379 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ms5g" event={"ID":"5e4632ba-3288-40e2-adf8-671fe25a869b","Type":"ContainerDied","Data":"e6591f36c50290a92d87c5edd46e3c0eff59a4fbe62b90cb88b3aae391353633"} Nov 27 11:06:21 crc kubenswrapper[4806]: I1127 11:06:21.035431 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ms5g" event={"ID":"5e4632ba-3288-40e2-adf8-671fe25a869b","Type":"ContainerStarted","Data":"d738ff0daa4894367b8a6700073ff5e91b03f31fb838962d231d7d726bd186ac"} Nov 27 11:06:21 crc kubenswrapper[4806]: I1127 11:06:21.038251 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 11:06:24 crc kubenswrapper[4806]: I1127 11:06:24.115395 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ms5g" event={"ID":"5e4632ba-3288-40e2-adf8-671fe25a869b","Type":"ContainerStarted","Data":"3f38b54e123fc9835c246e9d993a5e5b36e9d5954d68064c93927b459c1652e5"} Nov 27 11:06:26 crc kubenswrapper[4806]: I1127 11:06:26.149100 4806 generic.go:334] "Generic (PLEG): container finished" podID="5e4632ba-3288-40e2-adf8-671fe25a869b" containerID="3f38b54e123fc9835c246e9d993a5e5b36e9d5954d68064c93927b459c1652e5" exitCode=0 Nov 27 11:06:26 crc kubenswrapper[4806]: I1127 11:06:26.151295 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ms5g" event={"ID":"5e4632ba-3288-40e2-adf8-671fe25a869b","Type":"ContainerDied","Data":"3f38b54e123fc9835c246e9d993a5e5b36e9d5954d68064c93927b459c1652e5"} Nov 27 11:06:27 crc kubenswrapper[4806]: I1127 11:06:27.165626 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ms5g" event={"ID":"5e4632ba-3288-40e2-adf8-671fe25a869b","Type":"ContainerStarted","Data":"9700ac9eccc43fc0407bfcc362c350d4b0853f876bbf62873ac257763ad82354"} Nov 27 11:06:27 crc kubenswrapper[4806]: I1127 11:06:27.195464 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5ms5g" podStartSLOduration=2.570054126 podStartE2EDuration="8.195443833s" podCreationTimestamp="2025-11-27 11:06:19 +0000 UTC" firstStartedPulling="2025-11-27 11:06:21.037940664 +0000 UTC m=+2685.624531428" lastFinishedPulling="2025-11-27 11:06:26.663330371 +0000 UTC m=+2691.249921135" observedRunningTime="2025-11-27 11:06:27.194000293 +0000 UTC m=+2691.780591057" watchObservedRunningTime="2025-11-27 11:06:27.195443833 +0000 UTC m=+2691.782034587" Nov 27 11:06:29 crc kubenswrapper[4806]: I1127 11:06:29.955734 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:29 crc kubenswrapper[4806]: I1127 11:06:29.956035 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:30 crc kubenswrapper[4806]: I1127 11:06:30.017177 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:40 crc kubenswrapper[4806]: I1127 11:06:40.018137 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:40 crc kubenswrapper[4806]: I1127 11:06:40.091765 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ms5g"] Nov 27 11:06:40 crc kubenswrapper[4806]: I1127 11:06:40.320104 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5ms5g" podUID="5e4632ba-3288-40e2-adf8-671fe25a869b" containerName="registry-server" containerID="cri-o://9700ac9eccc43fc0407bfcc362c350d4b0853f876bbf62873ac257763ad82354" gracePeriod=2 Nov 27 11:06:40 crc kubenswrapper[4806]: I1127 11:06:40.783374 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:40 crc kubenswrapper[4806]: I1127 11:06:40.888568 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e4632ba-3288-40e2-adf8-671fe25a869b-catalog-content\") pod \"5e4632ba-3288-40e2-adf8-671fe25a869b\" (UID: \"5e4632ba-3288-40e2-adf8-671fe25a869b\") " Nov 27 11:06:40 crc kubenswrapper[4806]: I1127 11:06:40.888989 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcpfc\" (UniqueName: \"kubernetes.io/projected/5e4632ba-3288-40e2-adf8-671fe25a869b-kube-api-access-zcpfc\") pod \"5e4632ba-3288-40e2-adf8-671fe25a869b\" (UID: \"5e4632ba-3288-40e2-adf8-671fe25a869b\") " Nov 27 11:06:40 crc kubenswrapper[4806]: I1127 11:06:40.889053 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e4632ba-3288-40e2-adf8-671fe25a869b-utilities\") pod \"5e4632ba-3288-40e2-adf8-671fe25a869b\" (UID: \"5e4632ba-3288-40e2-adf8-671fe25a869b\") " Nov 27 11:06:40 crc kubenswrapper[4806]: I1127 11:06:40.890167 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e4632ba-3288-40e2-adf8-671fe25a869b-utilities" (OuterVolumeSpecName: "utilities") pod "5e4632ba-3288-40e2-adf8-671fe25a869b" (UID: "5e4632ba-3288-40e2-adf8-671fe25a869b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:06:40 crc kubenswrapper[4806]: I1127 11:06:40.915268 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e4632ba-3288-40e2-adf8-671fe25a869b-kube-api-access-zcpfc" (OuterVolumeSpecName: "kube-api-access-zcpfc") pod "5e4632ba-3288-40e2-adf8-671fe25a869b" (UID: "5e4632ba-3288-40e2-adf8-671fe25a869b"). InnerVolumeSpecName "kube-api-access-zcpfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:06:40 crc kubenswrapper[4806]: I1127 11:06:40.934391 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e4632ba-3288-40e2-adf8-671fe25a869b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e4632ba-3288-40e2-adf8-671fe25a869b" (UID: "5e4632ba-3288-40e2-adf8-671fe25a869b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:06:40 crc kubenswrapper[4806]: I1127 11:06:40.992657 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e4632ba-3288-40e2-adf8-671fe25a869b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:06:40 crc kubenswrapper[4806]: I1127 11:06:40.992709 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcpfc\" (UniqueName: \"kubernetes.io/projected/5e4632ba-3288-40e2-adf8-671fe25a869b-kube-api-access-zcpfc\") on node \"crc\" DevicePath \"\"" Nov 27 11:06:40 crc kubenswrapper[4806]: I1127 11:06:40.992782 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e4632ba-3288-40e2-adf8-671fe25a869b-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.331973 4806 generic.go:334] "Generic (PLEG): container finished" podID="5e4632ba-3288-40e2-adf8-671fe25a869b" containerID="9700ac9eccc43fc0407bfcc362c350d4b0853f876bbf62873ac257763ad82354" exitCode=0 Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.332059 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5ms5g" Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.332198 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ms5g" event={"ID":"5e4632ba-3288-40e2-adf8-671fe25a869b","Type":"ContainerDied","Data":"9700ac9eccc43fc0407bfcc362c350d4b0853f876bbf62873ac257763ad82354"} Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.332273 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5ms5g" event={"ID":"5e4632ba-3288-40e2-adf8-671fe25a869b","Type":"ContainerDied","Data":"d738ff0daa4894367b8a6700073ff5e91b03f31fb838962d231d7d726bd186ac"} Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.332299 4806 scope.go:117] "RemoveContainer" containerID="9700ac9eccc43fc0407bfcc362c350d4b0853f876bbf62873ac257763ad82354" Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.375388 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ms5g"] Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.377491 4806 scope.go:117] "RemoveContainer" containerID="3f38b54e123fc9835c246e9d993a5e5b36e9d5954d68064c93927b459c1652e5" Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.388274 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5ms5g"] Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.400276 4806 scope.go:117] "RemoveContainer" containerID="e6591f36c50290a92d87c5edd46e3c0eff59a4fbe62b90cb88b3aae391353633" Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.442205 4806 scope.go:117] "RemoveContainer" containerID="9700ac9eccc43fc0407bfcc362c350d4b0853f876bbf62873ac257763ad82354" Nov 27 11:06:41 crc kubenswrapper[4806]: E1127 11:06:41.442723 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9700ac9eccc43fc0407bfcc362c350d4b0853f876bbf62873ac257763ad82354\": container with ID starting with 9700ac9eccc43fc0407bfcc362c350d4b0853f876bbf62873ac257763ad82354 not found: ID does not exist" containerID="9700ac9eccc43fc0407bfcc362c350d4b0853f876bbf62873ac257763ad82354" Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.442765 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9700ac9eccc43fc0407bfcc362c350d4b0853f876bbf62873ac257763ad82354"} err="failed to get container status \"9700ac9eccc43fc0407bfcc362c350d4b0853f876bbf62873ac257763ad82354\": rpc error: code = NotFound desc = could not find container \"9700ac9eccc43fc0407bfcc362c350d4b0853f876bbf62873ac257763ad82354\": container with ID starting with 9700ac9eccc43fc0407bfcc362c350d4b0853f876bbf62873ac257763ad82354 not found: ID does not exist" Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.442792 4806 scope.go:117] "RemoveContainer" containerID="3f38b54e123fc9835c246e9d993a5e5b36e9d5954d68064c93927b459c1652e5" Nov 27 11:06:41 crc kubenswrapper[4806]: E1127 11:06:41.443124 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f38b54e123fc9835c246e9d993a5e5b36e9d5954d68064c93927b459c1652e5\": container with ID starting with 3f38b54e123fc9835c246e9d993a5e5b36e9d5954d68064c93927b459c1652e5 not found: ID does not exist" containerID="3f38b54e123fc9835c246e9d993a5e5b36e9d5954d68064c93927b459c1652e5" Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.443149 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f38b54e123fc9835c246e9d993a5e5b36e9d5954d68064c93927b459c1652e5"} err="failed to get container status \"3f38b54e123fc9835c246e9d993a5e5b36e9d5954d68064c93927b459c1652e5\": rpc error: code = NotFound desc = could not find container \"3f38b54e123fc9835c246e9d993a5e5b36e9d5954d68064c93927b459c1652e5\": container with ID starting with 3f38b54e123fc9835c246e9d993a5e5b36e9d5954d68064c93927b459c1652e5 not found: ID does not exist" Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.443166 4806 scope.go:117] "RemoveContainer" containerID="e6591f36c50290a92d87c5edd46e3c0eff59a4fbe62b90cb88b3aae391353633" Nov 27 11:06:41 crc kubenswrapper[4806]: E1127 11:06:41.443370 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6591f36c50290a92d87c5edd46e3c0eff59a4fbe62b90cb88b3aae391353633\": container with ID starting with e6591f36c50290a92d87c5edd46e3c0eff59a4fbe62b90cb88b3aae391353633 not found: ID does not exist" containerID="e6591f36c50290a92d87c5edd46e3c0eff59a4fbe62b90cb88b3aae391353633" Nov 27 11:06:41 crc kubenswrapper[4806]: I1127 11:06:41.443390 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6591f36c50290a92d87c5edd46e3c0eff59a4fbe62b90cb88b3aae391353633"} err="failed to get container status \"e6591f36c50290a92d87c5edd46e3c0eff59a4fbe62b90cb88b3aae391353633\": rpc error: code = NotFound desc = could not find container \"e6591f36c50290a92d87c5edd46e3c0eff59a4fbe62b90cb88b3aae391353633\": container with ID starting with e6591f36c50290a92d87c5edd46e3c0eff59a4fbe62b90cb88b3aae391353633 not found: ID does not exist" Nov 27 11:06:41 crc kubenswrapper[4806]: E1127 11:06:41.523222 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e4632ba_3288_40e2_adf8_671fe25a869b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e4632ba_3288_40e2_adf8_671fe25a869b.slice/crio-d738ff0daa4894367b8a6700073ff5e91b03f31fb838962d231d7d726bd186ac\": RecentStats: unable to find data in memory cache]" Nov 27 11:06:42 crc kubenswrapper[4806]: I1127 11:06:42.136212 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e4632ba-3288-40e2-adf8-671fe25a869b" path="/var/lib/kubelet/pods/5e4632ba-3288-40e2-adf8-671fe25a869b/volumes" Nov 27 11:06:44 crc kubenswrapper[4806]: I1127 11:06:44.450905 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:06:44 crc kubenswrapper[4806]: I1127 11:06:44.451225 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:06:44 crc kubenswrapper[4806]: I1127 11:06:44.451303 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 11:06:44 crc kubenswrapper[4806]: I1127 11:06:44.452143 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3a74cf9f736bfbb3f039de322c4fc66a3aea02cb5343e605f8777b301781b8fe"} pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 11:06:44 crc kubenswrapper[4806]: I1127 11:06:44.452246 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" containerID="cri-o://3a74cf9f736bfbb3f039de322c4fc66a3aea02cb5343e605f8777b301781b8fe" gracePeriod=600 Nov 27 11:06:45 crc kubenswrapper[4806]: I1127 11:06:45.382383 4806 generic.go:334] "Generic (PLEG): container finished" podID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerID="3a74cf9f736bfbb3f039de322c4fc66a3aea02cb5343e605f8777b301781b8fe" exitCode=0 Nov 27 11:06:45 crc kubenswrapper[4806]: I1127 11:06:45.382488 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerDied","Data":"3a74cf9f736bfbb3f039de322c4fc66a3aea02cb5343e605f8777b301781b8fe"} Nov 27 11:06:45 crc kubenswrapper[4806]: I1127 11:06:45.383275 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerStarted","Data":"3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4"} Nov 27 11:06:45 crc kubenswrapper[4806]: I1127 11:06:45.383299 4806 scope.go:117] "RemoveContainer" containerID="90fcd8c2c2d919e7dad1ab7f4ebe2e056a74fd74a0b5d71d38e04d60d6319ecc" Nov 27 11:06:49 crc kubenswrapper[4806]: I1127 11:06:49.433149 4806 generic.go:334] "Generic (PLEG): container finished" podID="d3e2d05c-f6bb-4072-a353-d811c8d67307" containerID="fa261d446bd2845eb5d06e7e61b4ad539b9c5b0ecf95406bdda373790e8c4475" exitCode=0 Nov 27 11:06:49 crc kubenswrapper[4806]: I1127 11:06:49.433886 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vbn8q/must-gather-7wk8j" event={"ID":"d3e2d05c-f6bb-4072-a353-d811c8d67307","Type":"ContainerDied","Data":"fa261d446bd2845eb5d06e7e61b4ad539b9c5b0ecf95406bdda373790e8c4475"} Nov 27 11:06:49 crc kubenswrapper[4806]: I1127 11:06:49.434706 4806 scope.go:117] "RemoveContainer" containerID="fa261d446bd2845eb5d06e7e61b4ad539b9c5b0ecf95406bdda373790e8c4475" Nov 27 11:06:49 crc kubenswrapper[4806]: I1127 11:06:49.516245 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vbn8q_must-gather-7wk8j_d3e2d05c-f6bb-4072-a353-d811c8d67307/gather/0.log" Nov 27 11:06:57 crc kubenswrapper[4806]: I1127 11:06:57.631978 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vbn8q/must-gather-7wk8j"] Nov 27 11:06:57 crc kubenswrapper[4806]: I1127 11:06:57.633223 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-vbn8q/must-gather-7wk8j" podUID="d3e2d05c-f6bb-4072-a353-d811c8d67307" containerName="copy" containerID="cri-o://40154098684ec914750b51d528275ceda1c92aec958d89187daa648e688e5f89" gracePeriod=2 Nov 27 11:06:57 crc kubenswrapper[4806]: I1127 11:06:57.650192 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vbn8q/must-gather-7wk8j"] Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.167352 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vbn8q_must-gather-7wk8j_d3e2d05c-f6bb-4072-a353-d811c8d67307/copy/0.log" Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.168775 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vbn8q/must-gather-7wk8j" Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.192431 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d3e2d05c-f6bb-4072-a353-d811c8d67307-must-gather-output\") pod \"d3e2d05c-f6bb-4072-a353-d811c8d67307\" (UID: \"d3e2d05c-f6bb-4072-a353-d811c8d67307\") " Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.192486 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdsbh\" (UniqueName: \"kubernetes.io/projected/d3e2d05c-f6bb-4072-a353-d811c8d67307-kube-api-access-zdsbh\") pod \"d3e2d05c-f6bb-4072-a353-d811c8d67307\" (UID: \"d3e2d05c-f6bb-4072-a353-d811c8d67307\") " Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.200491 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3e2d05c-f6bb-4072-a353-d811c8d67307-kube-api-access-zdsbh" (OuterVolumeSpecName: "kube-api-access-zdsbh") pod "d3e2d05c-f6bb-4072-a353-d811c8d67307" (UID: "d3e2d05c-f6bb-4072-a353-d811c8d67307"). InnerVolumeSpecName "kube-api-access-zdsbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.295224 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdsbh\" (UniqueName: \"kubernetes.io/projected/d3e2d05c-f6bb-4072-a353-d811c8d67307-kube-api-access-zdsbh\") on node \"crc\" DevicePath \"\"" Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.361419 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3e2d05c-f6bb-4072-a353-d811c8d67307-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d3e2d05c-f6bb-4072-a353-d811c8d67307" (UID: "d3e2d05c-f6bb-4072-a353-d811c8d67307"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.397940 4806 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d3e2d05c-f6bb-4072-a353-d811c8d67307-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.510133 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vbn8q_must-gather-7wk8j_d3e2d05c-f6bb-4072-a353-d811c8d67307/copy/0.log" Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.510810 4806 generic.go:334] "Generic (PLEG): container finished" podID="d3e2d05c-f6bb-4072-a353-d811c8d67307" containerID="40154098684ec914750b51d528275ceda1c92aec958d89187daa648e688e5f89" exitCode=143 Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.510878 4806 scope.go:117] "RemoveContainer" containerID="40154098684ec914750b51d528275ceda1c92aec958d89187daa648e688e5f89" Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.511071 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vbn8q/must-gather-7wk8j" Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.543089 4806 scope.go:117] "RemoveContainer" containerID="fa261d446bd2845eb5d06e7e61b4ad539b9c5b0ecf95406bdda373790e8c4475" Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.626564 4806 scope.go:117] "RemoveContainer" containerID="40154098684ec914750b51d528275ceda1c92aec958d89187daa648e688e5f89" Nov 27 11:06:58 crc kubenswrapper[4806]: E1127 11:06:58.627188 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40154098684ec914750b51d528275ceda1c92aec958d89187daa648e688e5f89\": container with ID starting with 40154098684ec914750b51d528275ceda1c92aec958d89187daa648e688e5f89 not found: ID does not exist" containerID="40154098684ec914750b51d528275ceda1c92aec958d89187daa648e688e5f89" Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.627243 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40154098684ec914750b51d528275ceda1c92aec958d89187daa648e688e5f89"} err="failed to get container status \"40154098684ec914750b51d528275ceda1c92aec958d89187daa648e688e5f89\": rpc error: code = NotFound desc = could not find container \"40154098684ec914750b51d528275ceda1c92aec958d89187daa648e688e5f89\": container with ID starting with 40154098684ec914750b51d528275ceda1c92aec958d89187daa648e688e5f89 not found: ID does not exist" Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.627270 4806 scope.go:117] "RemoveContainer" containerID="fa261d446bd2845eb5d06e7e61b4ad539b9c5b0ecf95406bdda373790e8c4475" Nov 27 11:06:58 crc kubenswrapper[4806]: E1127 11:06:58.627690 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa261d446bd2845eb5d06e7e61b4ad539b9c5b0ecf95406bdda373790e8c4475\": container with ID starting with fa261d446bd2845eb5d06e7e61b4ad539b9c5b0ecf95406bdda373790e8c4475 not found: ID does not exist" containerID="fa261d446bd2845eb5d06e7e61b4ad539b9c5b0ecf95406bdda373790e8c4475" Nov 27 11:06:58 crc kubenswrapper[4806]: I1127 11:06:58.627720 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa261d446bd2845eb5d06e7e61b4ad539b9c5b0ecf95406bdda373790e8c4475"} err="failed to get container status \"fa261d446bd2845eb5d06e7e61b4ad539b9c5b0ecf95406bdda373790e8c4475\": rpc error: code = NotFound desc = could not find container \"fa261d446bd2845eb5d06e7e61b4ad539b9c5b0ecf95406bdda373790e8c4475\": container with ID starting with fa261d446bd2845eb5d06e7e61b4ad539b9c5b0ecf95406bdda373790e8c4475 not found: ID does not exist" Nov 27 11:07:00 crc kubenswrapper[4806]: I1127 11:07:00.128897 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3e2d05c-f6bb-4072-a353-d811c8d67307" path="/var/lib/kubelet/pods/d3e2d05c-f6bb-4072-a353-d811c8d67307/volumes" Nov 27 11:08:38 crc kubenswrapper[4806]: I1127 11:08:38.821209 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zw4mm"] Nov 27 11:08:38 crc kubenswrapper[4806]: E1127 11:08:38.822163 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e2d05c-f6bb-4072-a353-d811c8d67307" containerName="copy" Nov 27 11:08:38 crc kubenswrapper[4806]: I1127 11:08:38.822189 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e2d05c-f6bb-4072-a353-d811c8d67307" containerName="copy" Nov 27 11:08:38 crc kubenswrapper[4806]: E1127 11:08:38.822210 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4632ba-3288-40e2-adf8-671fe25a869b" containerName="extract-content" Nov 27 11:08:38 crc kubenswrapper[4806]: I1127 11:08:38.822217 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4632ba-3288-40e2-adf8-671fe25a869b" containerName="extract-content" Nov 27 11:08:38 crc kubenswrapper[4806]: E1127 11:08:38.822239 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4632ba-3288-40e2-adf8-671fe25a869b" containerName="extract-utilities" Nov 27 11:08:38 crc kubenswrapper[4806]: I1127 11:08:38.822246 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4632ba-3288-40e2-adf8-671fe25a869b" containerName="extract-utilities" Nov 27 11:08:38 crc kubenswrapper[4806]: E1127 11:08:38.822263 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4632ba-3288-40e2-adf8-671fe25a869b" containerName="registry-server" Nov 27 11:08:38 crc kubenswrapper[4806]: I1127 11:08:38.822268 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4632ba-3288-40e2-adf8-671fe25a869b" containerName="registry-server" Nov 27 11:08:38 crc kubenswrapper[4806]: E1127 11:08:38.822281 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e2d05c-f6bb-4072-a353-d811c8d67307" containerName="gather" Nov 27 11:08:38 crc kubenswrapper[4806]: I1127 11:08:38.822288 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e2d05c-f6bb-4072-a353-d811c8d67307" containerName="gather" Nov 27 11:08:38 crc kubenswrapper[4806]: I1127 11:08:38.822467 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e2d05c-f6bb-4072-a353-d811c8d67307" containerName="copy" Nov 27 11:08:38 crc kubenswrapper[4806]: I1127 11:08:38.822494 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e2d05c-f6bb-4072-a353-d811c8d67307" containerName="gather" Nov 27 11:08:38 crc kubenswrapper[4806]: I1127 11:08:38.822505 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e4632ba-3288-40e2-adf8-671fe25a869b" containerName="registry-server" Nov 27 11:08:38 crc kubenswrapper[4806]: I1127 11:08:38.825578 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:38 crc kubenswrapper[4806]: I1127 11:08:38.879852 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zw4mm"] Nov 27 11:08:39 crc kubenswrapper[4806]: I1127 11:08:39.028250 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx9wd\" (UniqueName: \"kubernetes.io/projected/6ddc4e38-45e9-456c-a844-ac585afce8eb-kube-api-access-cx9wd\") pod \"certified-operators-zw4mm\" (UID: \"6ddc4e38-45e9-456c-a844-ac585afce8eb\") " pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:39 crc kubenswrapper[4806]: I1127 11:08:39.028763 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ddc4e38-45e9-456c-a844-ac585afce8eb-catalog-content\") pod \"certified-operators-zw4mm\" (UID: \"6ddc4e38-45e9-456c-a844-ac585afce8eb\") " pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:39 crc kubenswrapper[4806]: I1127 11:08:39.028818 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ddc4e38-45e9-456c-a844-ac585afce8eb-utilities\") pod \"certified-operators-zw4mm\" (UID: \"6ddc4e38-45e9-456c-a844-ac585afce8eb\") " pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:39 crc kubenswrapper[4806]: I1127 11:08:39.130793 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ddc4e38-45e9-456c-a844-ac585afce8eb-catalog-content\") pod \"certified-operators-zw4mm\" (UID: \"6ddc4e38-45e9-456c-a844-ac585afce8eb\") " pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:39 crc kubenswrapper[4806]: I1127 11:08:39.131031 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ddc4e38-45e9-456c-a844-ac585afce8eb-utilities\") pod \"certified-operators-zw4mm\" (UID: \"6ddc4e38-45e9-456c-a844-ac585afce8eb\") " pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:39 crc kubenswrapper[4806]: I1127 11:08:39.131212 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx9wd\" (UniqueName: \"kubernetes.io/projected/6ddc4e38-45e9-456c-a844-ac585afce8eb-kube-api-access-cx9wd\") pod \"certified-operators-zw4mm\" (UID: \"6ddc4e38-45e9-456c-a844-ac585afce8eb\") " pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:39 crc kubenswrapper[4806]: I1127 11:08:39.131582 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ddc4e38-45e9-456c-a844-ac585afce8eb-utilities\") pod \"certified-operators-zw4mm\" (UID: \"6ddc4e38-45e9-456c-a844-ac585afce8eb\") " pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:39 crc kubenswrapper[4806]: I1127 11:08:39.132071 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ddc4e38-45e9-456c-a844-ac585afce8eb-catalog-content\") pod \"certified-operators-zw4mm\" (UID: \"6ddc4e38-45e9-456c-a844-ac585afce8eb\") " pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:39 crc kubenswrapper[4806]: I1127 11:08:39.172493 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx9wd\" (UniqueName: \"kubernetes.io/projected/6ddc4e38-45e9-456c-a844-ac585afce8eb-kube-api-access-cx9wd\") pod \"certified-operators-zw4mm\" (UID: \"6ddc4e38-45e9-456c-a844-ac585afce8eb\") " pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:39 crc kubenswrapper[4806]: I1127 11:08:39.460101 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:39 crc kubenswrapper[4806]: I1127 11:08:39.967547 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zw4mm"] Nov 27 11:08:40 crc kubenswrapper[4806]: I1127 11:08:40.409316 4806 generic.go:334] "Generic (PLEG): container finished" podID="6ddc4e38-45e9-456c-a844-ac585afce8eb" containerID="f792e7ac769f451dc242506e39af071ab753972129b0ec81e0bb9bcc2e1f37e7" exitCode=0 Nov 27 11:08:40 crc kubenswrapper[4806]: I1127 11:08:40.409386 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4mm" event={"ID":"6ddc4e38-45e9-456c-a844-ac585afce8eb","Type":"ContainerDied","Data":"f792e7ac769f451dc242506e39af071ab753972129b0ec81e0bb9bcc2e1f37e7"} Nov 27 11:08:40 crc kubenswrapper[4806]: I1127 11:08:40.409671 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4mm" event={"ID":"6ddc4e38-45e9-456c-a844-ac585afce8eb","Type":"ContainerStarted","Data":"d4ce663fe470c905a46b27a6a034e32514bb2a8bcb5fde088bd48897f1224e35"} Nov 27 11:08:42 crc kubenswrapper[4806]: I1127 11:08:42.425341 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4mm" event={"ID":"6ddc4e38-45e9-456c-a844-ac585afce8eb","Type":"ContainerStarted","Data":"02235f8c2707325bcc054eb41c8f5f1cb8d402b8248b976891d21d3ad3b537e3"} Nov 27 11:08:43 crc kubenswrapper[4806]: I1127 11:08:43.439331 4806 generic.go:334] "Generic (PLEG): container finished" podID="6ddc4e38-45e9-456c-a844-ac585afce8eb" containerID="02235f8c2707325bcc054eb41c8f5f1cb8d402b8248b976891d21d3ad3b537e3" exitCode=0 Nov 27 11:08:43 crc kubenswrapper[4806]: I1127 11:08:43.439382 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4mm" event={"ID":"6ddc4e38-45e9-456c-a844-ac585afce8eb","Type":"ContainerDied","Data":"02235f8c2707325bcc054eb41c8f5f1cb8d402b8248b976891d21d3ad3b537e3"} Nov 27 11:08:43 crc kubenswrapper[4806]: I1127 11:08:43.794568 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p5pkn"] Nov 27 11:08:43 crc kubenswrapper[4806]: I1127 11:08:43.797543 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:43 crc kubenswrapper[4806]: I1127 11:08:43.807275 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p5pkn"] Nov 27 11:08:43 crc kubenswrapper[4806]: I1127 11:08:43.924938 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qnw2\" (UniqueName: \"kubernetes.io/projected/d2e010ac-83dc-42c0-a472-1957890989ff-kube-api-access-5qnw2\") pod \"community-operators-p5pkn\" (UID: \"d2e010ac-83dc-42c0-a472-1957890989ff\") " pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:43 crc kubenswrapper[4806]: I1127 11:08:43.925120 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2e010ac-83dc-42c0-a472-1957890989ff-utilities\") pod \"community-operators-p5pkn\" (UID: \"d2e010ac-83dc-42c0-a472-1957890989ff\") " pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:43 crc kubenswrapper[4806]: I1127 11:08:43.925149 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2e010ac-83dc-42c0-a472-1957890989ff-catalog-content\") pod \"community-operators-p5pkn\" (UID: \"d2e010ac-83dc-42c0-a472-1957890989ff\") " pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:44 crc kubenswrapper[4806]: I1127 11:08:44.026805 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2e010ac-83dc-42c0-a472-1957890989ff-utilities\") pod \"community-operators-p5pkn\" (UID: \"d2e010ac-83dc-42c0-a472-1957890989ff\") " pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:44 crc kubenswrapper[4806]: I1127 11:08:44.026855 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2e010ac-83dc-42c0-a472-1957890989ff-catalog-content\") pod \"community-operators-p5pkn\" (UID: \"d2e010ac-83dc-42c0-a472-1957890989ff\") " pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:44 crc kubenswrapper[4806]: I1127 11:08:44.026921 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qnw2\" (UniqueName: \"kubernetes.io/projected/d2e010ac-83dc-42c0-a472-1957890989ff-kube-api-access-5qnw2\") pod \"community-operators-p5pkn\" (UID: \"d2e010ac-83dc-42c0-a472-1957890989ff\") " pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:44 crc kubenswrapper[4806]: I1127 11:08:44.027363 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2e010ac-83dc-42c0-a472-1957890989ff-utilities\") pod \"community-operators-p5pkn\" (UID: \"d2e010ac-83dc-42c0-a472-1957890989ff\") " pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:44 crc kubenswrapper[4806]: I1127 11:08:44.027514 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2e010ac-83dc-42c0-a472-1957890989ff-catalog-content\") pod \"community-operators-p5pkn\" (UID: \"d2e010ac-83dc-42c0-a472-1957890989ff\") " pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:44 crc kubenswrapper[4806]: I1127 11:08:44.049399 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qnw2\" (UniqueName: \"kubernetes.io/projected/d2e010ac-83dc-42c0-a472-1957890989ff-kube-api-access-5qnw2\") pod \"community-operators-p5pkn\" (UID: \"d2e010ac-83dc-42c0-a472-1957890989ff\") " pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:44 crc kubenswrapper[4806]: I1127 11:08:44.122779 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:44 crc kubenswrapper[4806]: I1127 11:08:44.456800 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:08:44 crc kubenswrapper[4806]: I1127 11:08:44.457194 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:08:44 crc kubenswrapper[4806]: I1127 11:08:44.670409 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p5pkn"] Nov 27 11:08:45 crc kubenswrapper[4806]: I1127 11:08:45.486303 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4mm" event={"ID":"6ddc4e38-45e9-456c-a844-ac585afce8eb","Type":"ContainerStarted","Data":"a405ec1f411c3f39b7d9406627ed4c6ddb3d227998ee522e50451875150ab45c"} Nov 27 11:08:45 crc kubenswrapper[4806]: I1127 11:08:45.489142 4806 generic.go:334] "Generic (PLEG): container finished" podID="d2e010ac-83dc-42c0-a472-1957890989ff" containerID="3d95a333cf04411da4a88f06289a025ce177d9868cfc0b4c3e9772c5257cdea4" exitCode=0 Nov 27 11:08:45 crc kubenswrapper[4806]: I1127 11:08:45.489194 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5pkn" event={"ID":"d2e010ac-83dc-42c0-a472-1957890989ff","Type":"ContainerDied","Data":"3d95a333cf04411da4a88f06289a025ce177d9868cfc0b4c3e9772c5257cdea4"} Nov 27 11:08:45 crc kubenswrapper[4806]: I1127 11:08:45.489243 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5pkn" event={"ID":"d2e010ac-83dc-42c0-a472-1957890989ff","Type":"ContainerStarted","Data":"68ca28ad2430659a0b3bf44726957ff473fdd56d1bd54ad402f2a6f36d36989b"} Nov 27 11:08:45 crc kubenswrapper[4806]: I1127 11:08:45.522460 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zw4mm" podStartSLOduration=3.7476906789999997 podStartE2EDuration="7.522437098s" podCreationTimestamp="2025-11-27 11:08:38 +0000 UTC" firstStartedPulling="2025-11-27 11:08:40.411399212 +0000 UTC m=+2824.997989976" lastFinishedPulling="2025-11-27 11:08:44.186145631 +0000 UTC m=+2828.772736395" observedRunningTime="2025-11-27 11:08:45.514416676 +0000 UTC m=+2830.101007460" watchObservedRunningTime="2025-11-27 11:08:45.522437098 +0000 UTC m=+2830.109027862" Nov 27 11:08:47 crc kubenswrapper[4806]: I1127 11:08:47.506509 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5pkn" event={"ID":"d2e010ac-83dc-42c0-a472-1957890989ff","Type":"ContainerStarted","Data":"509ef7c56c9dec4e08733c44873dc9006ce81d15d032d87a7996b40304515004"} Nov 27 11:08:48 crc kubenswrapper[4806]: I1127 11:08:48.520902 4806 generic.go:334] "Generic (PLEG): container finished" podID="d2e010ac-83dc-42c0-a472-1957890989ff" containerID="509ef7c56c9dec4e08733c44873dc9006ce81d15d032d87a7996b40304515004" exitCode=0 Nov 27 11:08:48 crc kubenswrapper[4806]: I1127 11:08:48.520996 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5pkn" event={"ID":"d2e010ac-83dc-42c0-a472-1957890989ff","Type":"ContainerDied","Data":"509ef7c56c9dec4e08733c44873dc9006ce81d15d032d87a7996b40304515004"} Nov 27 11:08:49 crc kubenswrapper[4806]: I1127 11:08:49.460608 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:49 crc kubenswrapper[4806]: I1127 11:08:49.460938 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:49 crc kubenswrapper[4806]: I1127 11:08:49.507513 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:49 crc kubenswrapper[4806]: I1127 11:08:49.538366 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5pkn" event={"ID":"d2e010ac-83dc-42c0-a472-1957890989ff","Type":"ContainerStarted","Data":"744d966b355e782919c433b72095916283c5e63c7b9b05f46231a98fbde65528"} Nov 27 11:08:49 crc kubenswrapper[4806]: I1127 11:08:49.570605 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p5pkn" podStartSLOduration=3.02484756 podStartE2EDuration="6.570583886s" podCreationTimestamp="2025-11-27 11:08:43 +0000 UTC" firstStartedPulling="2025-11-27 11:08:45.491476802 +0000 UTC m=+2830.078067566" lastFinishedPulling="2025-11-27 11:08:49.037213128 +0000 UTC m=+2833.623803892" observedRunningTime="2025-11-27 11:08:49.564800646 +0000 UTC m=+2834.151391420" watchObservedRunningTime="2025-11-27 11:08:49.570583886 +0000 UTC m=+2834.157174650" Nov 27 11:08:49 crc kubenswrapper[4806]: I1127 11:08:49.591404 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:51 crc kubenswrapper[4806]: I1127 11:08:51.796400 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zw4mm"] Nov 27 11:08:51 crc kubenswrapper[4806]: I1127 11:08:51.798104 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zw4mm" podUID="6ddc4e38-45e9-456c-a844-ac585afce8eb" containerName="registry-server" containerID="cri-o://a405ec1f411c3f39b7d9406627ed4c6ddb3d227998ee522e50451875150ab45c" gracePeriod=2 Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.319680 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.513554 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ddc4e38-45e9-456c-a844-ac585afce8eb-utilities\") pod \"6ddc4e38-45e9-456c-a844-ac585afce8eb\" (UID: \"6ddc4e38-45e9-456c-a844-ac585afce8eb\") " Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.513622 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ddc4e38-45e9-456c-a844-ac585afce8eb-catalog-content\") pod \"6ddc4e38-45e9-456c-a844-ac585afce8eb\" (UID: \"6ddc4e38-45e9-456c-a844-ac585afce8eb\") " Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.513688 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx9wd\" (UniqueName: \"kubernetes.io/projected/6ddc4e38-45e9-456c-a844-ac585afce8eb-kube-api-access-cx9wd\") pod \"6ddc4e38-45e9-456c-a844-ac585afce8eb\" (UID: \"6ddc4e38-45e9-456c-a844-ac585afce8eb\") " Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.514813 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ddc4e38-45e9-456c-a844-ac585afce8eb-utilities" (OuterVolumeSpecName: "utilities") pod "6ddc4e38-45e9-456c-a844-ac585afce8eb" (UID: "6ddc4e38-45e9-456c-a844-ac585afce8eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.521579 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ddc4e38-45e9-456c-a844-ac585afce8eb-kube-api-access-cx9wd" (OuterVolumeSpecName: "kube-api-access-cx9wd") pod "6ddc4e38-45e9-456c-a844-ac585afce8eb" (UID: "6ddc4e38-45e9-456c-a844-ac585afce8eb"). InnerVolumeSpecName "kube-api-access-cx9wd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.571360 4806 generic.go:334] "Generic (PLEG): container finished" podID="6ddc4e38-45e9-456c-a844-ac585afce8eb" containerID="a405ec1f411c3f39b7d9406627ed4c6ddb3d227998ee522e50451875150ab45c" exitCode=0 Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.571421 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4mm" event={"ID":"6ddc4e38-45e9-456c-a844-ac585afce8eb","Type":"ContainerDied","Data":"a405ec1f411c3f39b7d9406627ed4c6ddb3d227998ee522e50451875150ab45c"} Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.571462 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zw4mm" event={"ID":"6ddc4e38-45e9-456c-a844-ac585afce8eb","Type":"ContainerDied","Data":"d4ce663fe470c905a46b27a6a034e32514bb2a8bcb5fde088bd48897f1224e35"} Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.571464 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zw4mm" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.571485 4806 scope.go:117] "RemoveContainer" containerID="a405ec1f411c3f39b7d9406627ed4c6ddb3d227998ee522e50451875150ab45c" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.596108 4806 scope.go:117] "RemoveContainer" containerID="02235f8c2707325bcc054eb41c8f5f1cb8d402b8248b976891d21d3ad3b537e3" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.617493 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ddc4e38-45e9-456c-a844-ac585afce8eb-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.617546 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx9wd\" (UniqueName: \"kubernetes.io/projected/6ddc4e38-45e9-456c-a844-ac585afce8eb-kube-api-access-cx9wd\") on node \"crc\" DevicePath \"\"" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.617752 4806 scope.go:117] "RemoveContainer" containerID="f792e7ac769f451dc242506e39af071ab753972129b0ec81e0bb9bcc2e1f37e7" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.660073 4806 scope.go:117] "RemoveContainer" containerID="a405ec1f411c3f39b7d9406627ed4c6ddb3d227998ee522e50451875150ab45c" Nov 27 11:08:52 crc kubenswrapper[4806]: E1127 11:08:52.660682 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a405ec1f411c3f39b7d9406627ed4c6ddb3d227998ee522e50451875150ab45c\": container with ID starting with a405ec1f411c3f39b7d9406627ed4c6ddb3d227998ee522e50451875150ab45c not found: ID does not exist" containerID="a405ec1f411c3f39b7d9406627ed4c6ddb3d227998ee522e50451875150ab45c" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.660719 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a405ec1f411c3f39b7d9406627ed4c6ddb3d227998ee522e50451875150ab45c"} err="failed to get container status \"a405ec1f411c3f39b7d9406627ed4c6ddb3d227998ee522e50451875150ab45c\": rpc error: code = NotFound desc = could not find container \"a405ec1f411c3f39b7d9406627ed4c6ddb3d227998ee522e50451875150ab45c\": container with ID starting with a405ec1f411c3f39b7d9406627ed4c6ddb3d227998ee522e50451875150ab45c not found: ID does not exist" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.660745 4806 scope.go:117] "RemoveContainer" containerID="02235f8c2707325bcc054eb41c8f5f1cb8d402b8248b976891d21d3ad3b537e3" Nov 27 11:08:52 crc kubenswrapper[4806]: E1127 11:08:52.661065 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02235f8c2707325bcc054eb41c8f5f1cb8d402b8248b976891d21d3ad3b537e3\": container with ID starting with 02235f8c2707325bcc054eb41c8f5f1cb8d402b8248b976891d21d3ad3b537e3 not found: ID does not exist" containerID="02235f8c2707325bcc054eb41c8f5f1cb8d402b8248b976891d21d3ad3b537e3" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.661102 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02235f8c2707325bcc054eb41c8f5f1cb8d402b8248b976891d21d3ad3b537e3"} err="failed to get container status \"02235f8c2707325bcc054eb41c8f5f1cb8d402b8248b976891d21d3ad3b537e3\": rpc error: code = NotFound desc = could not find container \"02235f8c2707325bcc054eb41c8f5f1cb8d402b8248b976891d21d3ad3b537e3\": container with ID starting with 02235f8c2707325bcc054eb41c8f5f1cb8d402b8248b976891d21d3ad3b537e3 not found: ID does not exist" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.661124 4806 scope.go:117] "RemoveContainer" containerID="f792e7ac769f451dc242506e39af071ab753972129b0ec81e0bb9bcc2e1f37e7" Nov 27 11:08:52 crc kubenswrapper[4806]: E1127 11:08:52.661435 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f792e7ac769f451dc242506e39af071ab753972129b0ec81e0bb9bcc2e1f37e7\": container with ID starting with f792e7ac769f451dc242506e39af071ab753972129b0ec81e0bb9bcc2e1f37e7 not found: ID does not exist" containerID="f792e7ac769f451dc242506e39af071ab753972129b0ec81e0bb9bcc2e1f37e7" Nov 27 11:08:52 crc kubenswrapper[4806]: I1127 11:08:52.661485 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f792e7ac769f451dc242506e39af071ab753972129b0ec81e0bb9bcc2e1f37e7"} err="failed to get container status \"f792e7ac769f451dc242506e39af071ab753972129b0ec81e0bb9bcc2e1f37e7\": rpc error: code = NotFound desc = could not find container \"f792e7ac769f451dc242506e39af071ab753972129b0ec81e0bb9bcc2e1f37e7\": container with ID starting with f792e7ac769f451dc242506e39af071ab753972129b0ec81e0bb9bcc2e1f37e7 not found: ID does not exist" Nov 27 11:08:53 crc kubenswrapper[4806]: I1127 11:08:53.268650 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ddc4e38-45e9-456c-a844-ac585afce8eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ddc4e38-45e9-456c-a844-ac585afce8eb" (UID: "6ddc4e38-45e9-456c-a844-ac585afce8eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:08:53 crc kubenswrapper[4806]: I1127 11:08:53.333190 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ddc4e38-45e9-456c-a844-ac585afce8eb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:08:53 crc kubenswrapper[4806]: I1127 11:08:53.519348 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zw4mm"] Nov 27 11:08:53 crc kubenswrapper[4806]: I1127 11:08:53.530183 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zw4mm"] Nov 27 11:08:54 crc kubenswrapper[4806]: I1127 11:08:54.132116 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ddc4e38-45e9-456c-a844-ac585afce8eb" path="/var/lib/kubelet/pods/6ddc4e38-45e9-456c-a844-ac585afce8eb/volumes" Nov 27 11:08:54 crc kubenswrapper[4806]: I1127 11:08:54.134069 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:54 crc kubenswrapper[4806]: I1127 11:08:54.134119 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:54 crc kubenswrapper[4806]: I1127 11:08:54.182933 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:54 crc kubenswrapper[4806]: I1127 11:08:54.657035 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:56 crc kubenswrapper[4806]: I1127 11:08:56.586813 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p5pkn"] Nov 27 11:08:56 crc kubenswrapper[4806]: I1127 11:08:56.612963 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p5pkn" podUID="d2e010ac-83dc-42c0-a472-1957890989ff" containerName="registry-server" containerID="cri-o://744d966b355e782919c433b72095916283c5e63c7b9b05f46231a98fbde65528" gracePeriod=2 Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.009667 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.129175 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2e010ac-83dc-42c0-a472-1957890989ff-catalog-content\") pod \"d2e010ac-83dc-42c0-a472-1957890989ff\" (UID: \"d2e010ac-83dc-42c0-a472-1957890989ff\") " Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.129365 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qnw2\" (UniqueName: \"kubernetes.io/projected/d2e010ac-83dc-42c0-a472-1957890989ff-kube-api-access-5qnw2\") pod \"d2e010ac-83dc-42c0-a472-1957890989ff\" (UID: \"d2e010ac-83dc-42c0-a472-1957890989ff\") " Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.129460 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2e010ac-83dc-42c0-a472-1957890989ff-utilities\") pod \"d2e010ac-83dc-42c0-a472-1957890989ff\" (UID: \"d2e010ac-83dc-42c0-a472-1957890989ff\") " Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.130642 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2e010ac-83dc-42c0-a472-1957890989ff-utilities" (OuterVolumeSpecName: "utilities") pod "d2e010ac-83dc-42c0-a472-1957890989ff" (UID: "d2e010ac-83dc-42c0-a472-1957890989ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.134991 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e010ac-83dc-42c0-a472-1957890989ff-kube-api-access-5qnw2" (OuterVolumeSpecName: "kube-api-access-5qnw2") pod "d2e010ac-83dc-42c0-a472-1957890989ff" (UID: "d2e010ac-83dc-42c0-a472-1957890989ff"). InnerVolumeSpecName "kube-api-access-5qnw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.184416 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2e010ac-83dc-42c0-a472-1957890989ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2e010ac-83dc-42c0-a472-1957890989ff" (UID: "d2e010ac-83dc-42c0-a472-1957890989ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.232981 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qnw2\" (UniqueName: \"kubernetes.io/projected/d2e010ac-83dc-42c0-a472-1957890989ff-kube-api-access-5qnw2\") on node \"crc\" DevicePath \"\"" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.233043 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2e010ac-83dc-42c0-a472-1957890989ff-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.233053 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2e010ac-83dc-42c0-a472-1957890989ff-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.626281 4806 generic.go:334] "Generic (PLEG): container finished" podID="d2e010ac-83dc-42c0-a472-1957890989ff" containerID="744d966b355e782919c433b72095916283c5e63c7b9b05f46231a98fbde65528" exitCode=0 Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.626376 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5pkn" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.626420 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5pkn" event={"ID":"d2e010ac-83dc-42c0-a472-1957890989ff","Type":"ContainerDied","Data":"744d966b355e782919c433b72095916283c5e63c7b9b05f46231a98fbde65528"} Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.626779 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5pkn" event={"ID":"d2e010ac-83dc-42c0-a472-1957890989ff","Type":"ContainerDied","Data":"68ca28ad2430659a0b3bf44726957ff473fdd56d1bd54ad402f2a6f36d36989b"} Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.626804 4806 scope.go:117] "RemoveContainer" containerID="744d966b355e782919c433b72095916283c5e63c7b9b05f46231a98fbde65528" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.680069 4806 scope.go:117] "RemoveContainer" containerID="509ef7c56c9dec4e08733c44873dc9006ce81d15d032d87a7996b40304515004" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.680502 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p5pkn"] Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.690185 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p5pkn"] Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.706478 4806 scope.go:117] "RemoveContainer" containerID="3d95a333cf04411da4a88f06289a025ce177d9868cfc0b4c3e9772c5257cdea4" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.773551 4806 scope.go:117] "RemoveContainer" containerID="744d966b355e782919c433b72095916283c5e63c7b9b05f46231a98fbde65528" Nov 27 11:08:57 crc kubenswrapper[4806]: E1127 11:08:57.777131 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"744d966b355e782919c433b72095916283c5e63c7b9b05f46231a98fbde65528\": container with ID starting with 744d966b355e782919c433b72095916283c5e63c7b9b05f46231a98fbde65528 not found: ID does not exist" containerID="744d966b355e782919c433b72095916283c5e63c7b9b05f46231a98fbde65528" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.777205 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"744d966b355e782919c433b72095916283c5e63c7b9b05f46231a98fbde65528"} err="failed to get container status \"744d966b355e782919c433b72095916283c5e63c7b9b05f46231a98fbde65528\": rpc error: code = NotFound desc = could not find container \"744d966b355e782919c433b72095916283c5e63c7b9b05f46231a98fbde65528\": container with ID starting with 744d966b355e782919c433b72095916283c5e63c7b9b05f46231a98fbde65528 not found: ID does not exist" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.777269 4806 scope.go:117] "RemoveContainer" containerID="509ef7c56c9dec4e08733c44873dc9006ce81d15d032d87a7996b40304515004" Nov 27 11:08:57 crc kubenswrapper[4806]: E1127 11:08:57.777568 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"509ef7c56c9dec4e08733c44873dc9006ce81d15d032d87a7996b40304515004\": container with ID starting with 509ef7c56c9dec4e08733c44873dc9006ce81d15d032d87a7996b40304515004 not found: ID does not exist" containerID="509ef7c56c9dec4e08733c44873dc9006ce81d15d032d87a7996b40304515004" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.777641 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"509ef7c56c9dec4e08733c44873dc9006ce81d15d032d87a7996b40304515004"} err="failed to get container status \"509ef7c56c9dec4e08733c44873dc9006ce81d15d032d87a7996b40304515004\": rpc error: code = NotFound desc = could not find container \"509ef7c56c9dec4e08733c44873dc9006ce81d15d032d87a7996b40304515004\": container with ID starting with 509ef7c56c9dec4e08733c44873dc9006ce81d15d032d87a7996b40304515004 not found: ID does not exist" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.777682 4806 scope.go:117] "RemoveContainer" containerID="3d95a333cf04411da4a88f06289a025ce177d9868cfc0b4c3e9772c5257cdea4" Nov 27 11:08:57 crc kubenswrapper[4806]: E1127 11:08:57.781420 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d95a333cf04411da4a88f06289a025ce177d9868cfc0b4c3e9772c5257cdea4\": container with ID starting with 3d95a333cf04411da4a88f06289a025ce177d9868cfc0b4c3e9772c5257cdea4 not found: ID does not exist" containerID="3d95a333cf04411da4a88f06289a025ce177d9868cfc0b4c3e9772c5257cdea4" Nov 27 11:08:57 crc kubenswrapper[4806]: I1127 11:08:57.781469 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d95a333cf04411da4a88f06289a025ce177d9868cfc0b4c3e9772c5257cdea4"} err="failed to get container status \"3d95a333cf04411da4a88f06289a025ce177d9868cfc0b4c3e9772c5257cdea4\": rpc error: code = NotFound desc = could not find container \"3d95a333cf04411da4a88f06289a025ce177d9868cfc0b4c3e9772c5257cdea4\": container with ID starting with 3d95a333cf04411da4a88f06289a025ce177d9868cfc0b4c3e9772c5257cdea4 not found: ID does not exist" Nov 27 11:08:58 crc kubenswrapper[4806]: I1127 11:08:58.129662 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2e010ac-83dc-42c0-a472-1957890989ff" path="/var/lib/kubelet/pods/d2e010ac-83dc-42c0-a472-1957890989ff/volumes" Nov 27 11:09:14 crc kubenswrapper[4806]: I1127 11:09:14.451201 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:09:14 crc kubenswrapper[4806]: I1127 11:09:14.451763 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:09:44 crc kubenswrapper[4806]: I1127 11:09:44.451197 4806 patch_prober.go:28] interesting pod/machine-config-daemon-7mq9p container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 27 11:09:44 crc kubenswrapper[4806]: I1127 11:09:44.451828 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 27 11:09:44 crc kubenswrapper[4806]: I1127 11:09:44.451896 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" Nov 27 11:09:44 crc kubenswrapper[4806]: I1127 11:09:44.452943 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4"} pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 27 11:09:44 crc kubenswrapper[4806]: I1127 11:09:44.453019 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerName="machine-config-daemon" containerID="cri-o://3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" gracePeriod=600 Nov 27 11:09:44 crc kubenswrapper[4806]: E1127 11:09:44.582316 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:09:45 crc kubenswrapper[4806]: I1127 11:09:45.265619 4806 generic.go:334] "Generic (PLEG): container finished" podID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" containerID="3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" exitCode=0 Nov 27 11:09:45 crc kubenswrapper[4806]: I1127 11:09:45.265657 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" event={"ID":"4ea0cad1-8e8b-458c-a943-7191b36e35a3","Type":"ContainerDied","Data":"3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4"} Nov 27 11:09:45 crc kubenswrapper[4806]: I1127 11:09:45.265747 4806 scope.go:117] "RemoveContainer" containerID="3a74cf9f736bfbb3f039de322c4fc66a3aea02cb5343e605f8777b301781b8fe" Nov 27 11:09:45 crc kubenswrapper[4806]: I1127 11:09:45.266452 4806 scope.go:117] "RemoveContainer" containerID="3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" Nov 27 11:09:45 crc kubenswrapper[4806]: E1127 11:09:45.266907 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:09:58 crc kubenswrapper[4806]: I1127 11:09:58.119939 4806 scope.go:117] "RemoveContainer" containerID="3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" Nov 27 11:09:58 crc kubenswrapper[4806]: E1127 11:09:58.122073 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:10:11 crc kubenswrapper[4806]: I1127 11:10:11.119486 4806 scope.go:117] "RemoveContainer" containerID="3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" Nov 27 11:10:11 crc kubenswrapper[4806]: E1127 11:10:11.120169 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:10:23 crc kubenswrapper[4806]: I1127 11:10:23.119114 4806 scope.go:117] "RemoveContainer" containerID="3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" Nov 27 11:10:23 crc kubenswrapper[4806]: E1127 11:10:23.120116 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:10:35 crc kubenswrapper[4806]: I1127 11:10:35.117789 4806 scope.go:117] "RemoveContainer" containerID="3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" Nov 27 11:10:35 crc kubenswrapper[4806]: E1127 11:10:35.121593 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:10:47 crc kubenswrapper[4806]: I1127 11:10:47.116780 4806 scope.go:117] "RemoveContainer" containerID="3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" Nov 27 11:10:47 crc kubenswrapper[4806]: E1127 11:10:47.117941 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:10:58 crc kubenswrapper[4806]: I1127 11:10:58.117499 4806 scope.go:117] "RemoveContainer" containerID="3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" Nov 27 11:10:58 crc kubenswrapper[4806]: E1127 11:10:58.119681 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:11:10 crc kubenswrapper[4806]: I1127 11:11:10.117984 4806 scope.go:117] "RemoveContainer" containerID="3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" Nov 27 11:11:10 crc kubenswrapper[4806]: E1127 11:11:10.119846 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.213897 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-66l69"] Nov 27 11:11:14 crc kubenswrapper[4806]: E1127 11:11:14.215397 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e010ac-83dc-42c0-a472-1957890989ff" containerName="registry-server" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.215413 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e010ac-83dc-42c0-a472-1957890989ff" containerName="registry-server" Nov 27 11:11:14 crc kubenswrapper[4806]: E1127 11:11:14.215427 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ddc4e38-45e9-456c-a844-ac585afce8eb" containerName="registry-server" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.215433 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ddc4e38-45e9-456c-a844-ac585afce8eb" containerName="registry-server" Nov 27 11:11:14 crc kubenswrapper[4806]: E1127 11:11:14.215458 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e010ac-83dc-42c0-a472-1957890989ff" containerName="extract-utilities" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.215465 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e010ac-83dc-42c0-a472-1957890989ff" containerName="extract-utilities" Nov 27 11:11:14 crc kubenswrapper[4806]: E1127 11:11:14.215473 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ddc4e38-45e9-456c-a844-ac585afce8eb" containerName="extract-utilities" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.215479 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ddc4e38-45e9-456c-a844-ac585afce8eb" containerName="extract-utilities" Nov 27 11:11:14 crc kubenswrapper[4806]: E1127 11:11:14.215495 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e010ac-83dc-42c0-a472-1957890989ff" containerName="extract-content" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.215501 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e010ac-83dc-42c0-a472-1957890989ff" containerName="extract-content" Nov 27 11:11:14 crc kubenswrapper[4806]: E1127 11:11:14.215512 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ddc4e38-45e9-456c-a844-ac585afce8eb" containerName="extract-content" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.215518 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ddc4e38-45e9-456c-a844-ac585afce8eb" containerName="extract-content" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.215738 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ddc4e38-45e9-456c-a844-ac585afce8eb" containerName="registry-server" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.215770 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e010ac-83dc-42c0-a472-1957890989ff" containerName="registry-server" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.217389 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.234224 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-66l69"] Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.357091 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88723dc1-3227-4511-89b9-62629792efee-utilities\") pod \"redhat-operators-66l69\" (UID: \"88723dc1-3227-4511-89b9-62629792efee\") " pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.357154 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88723dc1-3227-4511-89b9-62629792efee-catalog-content\") pod \"redhat-operators-66l69\" (UID: \"88723dc1-3227-4511-89b9-62629792efee\") " pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.357309 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bttbm\" (UniqueName: \"kubernetes.io/projected/88723dc1-3227-4511-89b9-62629792efee-kube-api-access-bttbm\") pod \"redhat-operators-66l69\" (UID: \"88723dc1-3227-4511-89b9-62629792efee\") " pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.459042 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bttbm\" (UniqueName: \"kubernetes.io/projected/88723dc1-3227-4511-89b9-62629792efee-kube-api-access-bttbm\") pod \"redhat-operators-66l69\" (UID: \"88723dc1-3227-4511-89b9-62629792efee\") " pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.459109 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88723dc1-3227-4511-89b9-62629792efee-utilities\") pod \"redhat-operators-66l69\" (UID: \"88723dc1-3227-4511-89b9-62629792efee\") " pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.459137 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88723dc1-3227-4511-89b9-62629792efee-catalog-content\") pod \"redhat-operators-66l69\" (UID: \"88723dc1-3227-4511-89b9-62629792efee\") " pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.459673 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88723dc1-3227-4511-89b9-62629792efee-catalog-content\") pod \"redhat-operators-66l69\" (UID: \"88723dc1-3227-4511-89b9-62629792efee\") " pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.459972 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88723dc1-3227-4511-89b9-62629792efee-utilities\") pod \"redhat-operators-66l69\" (UID: \"88723dc1-3227-4511-89b9-62629792efee\") " pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.485182 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bttbm\" (UniqueName: \"kubernetes.io/projected/88723dc1-3227-4511-89b9-62629792efee-kube-api-access-bttbm\") pod \"redhat-operators-66l69\" (UID: \"88723dc1-3227-4511-89b9-62629792efee\") " pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:14 crc kubenswrapper[4806]: I1127 11:11:14.540065 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:15 crc kubenswrapper[4806]: I1127 11:11:15.070788 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-66l69"] Nov 27 11:11:15 crc kubenswrapper[4806]: I1127 11:11:15.188757 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66l69" event={"ID":"88723dc1-3227-4511-89b9-62629792efee","Type":"ContainerStarted","Data":"dbb2c5de440d50a7270bf32a68e9d1318d7c6ff1f88fd9e4e8415ad50f31bbaa"} Nov 27 11:11:16 crc kubenswrapper[4806]: I1127 11:11:16.201438 4806 generic.go:334] "Generic (PLEG): container finished" podID="88723dc1-3227-4511-89b9-62629792efee" containerID="8d61b2293d73b991af200ec07e333bc664beaf9a4e90c8ee66b2356ae8e98b82" exitCode=0 Nov 27 11:11:16 crc kubenswrapper[4806]: I1127 11:11:16.201504 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66l69" event={"ID":"88723dc1-3227-4511-89b9-62629792efee","Type":"ContainerDied","Data":"8d61b2293d73b991af200ec07e333bc664beaf9a4e90c8ee66b2356ae8e98b82"} Nov 27 11:11:17 crc kubenswrapper[4806]: I1127 11:11:17.224087 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66l69" event={"ID":"88723dc1-3227-4511-89b9-62629792efee","Type":"ContainerStarted","Data":"24e14d0279dbb633f3aa65f40cc31f24609f9d33ddcb4241e0e1985dc334bd5a"} Nov 27 11:11:21 crc kubenswrapper[4806]: I1127 11:11:21.272223 4806 generic.go:334] "Generic (PLEG): container finished" podID="88723dc1-3227-4511-89b9-62629792efee" containerID="24e14d0279dbb633f3aa65f40cc31f24609f9d33ddcb4241e0e1985dc334bd5a" exitCode=0 Nov 27 11:11:21 crc kubenswrapper[4806]: I1127 11:11:21.273141 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66l69" event={"ID":"88723dc1-3227-4511-89b9-62629792efee","Type":"ContainerDied","Data":"24e14d0279dbb633f3aa65f40cc31f24609f9d33ddcb4241e0e1985dc334bd5a"} Nov 27 11:11:21 crc kubenswrapper[4806]: I1127 11:11:21.276877 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 27 11:11:22 crc kubenswrapper[4806]: I1127 11:11:22.285715 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66l69" event={"ID":"88723dc1-3227-4511-89b9-62629792efee","Type":"ContainerStarted","Data":"914d9d8d7a1697743b5bff33786c366f0996eb0ed9f378eb8569bd5c0a7d9812"} Nov 27 11:11:22 crc kubenswrapper[4806]: I1127 11:11:22.320997 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-66l69" podStartSLOduration=2.720279346 podStartE2EDuration="8.320967819s" podCreationTimestamp="2025-11-27 11:11:14 +0000 UTC" firstStartedPulling="2025-11-27 11:11:16.204430615 +0000 UTC m=+2980.791021379" lastFinishedPulling="2025-11-27 11:11:21.805119038 +0000 UTC m=+2986.391709852" observedRunningTime="2025-11-27 11:11:22.3063217 +0000 UTC m=+2986.892912504" watchObservedRunningTime="2025-11-27 11:11:22.320967819 +0000 UTC m=+2986.907558583" Nov 27 11:11:24 crc kubenswrapper[4806]: I1127 11:11:24.540421 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:24 crc kubenswrapper[4806]: I1127 11:11:24.542610 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:26 crc kubenswrapper[4806]: I1127 11:11:26.163798 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-66l69" podUID="88723dc1-3227-4511-89b9-62629792efee" containerName="registry-server" probeResult="failure" output=< Nov 27 11:11:26 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Nov 27 11:11:26 crc kubenswrapper[4806]: > Nov 27 11:11:26 crc kubenswrapper[4806]: I1127 11:11:26.191866 4806 scope.go:117] "RemoveContainer" containerID="3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" Nov 27 11:11:26 crc kubenswrapper[4806]: E1127 11:11:26.192371 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:11:34 crc kubenswrapper[4806]: I1127 11:11:34.635137 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:34 crc kubenswrapper[4806]: I1127 11:11:34.723733 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:34 crc kubenswrapper[4806]: I1127 11:11:34.893955 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-66l69"] Nov 27 11:11:36 crc kubenswrapper[4806]: I1127 11:11:36.358508 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-66l69" podUID="88723dc1-3227-4511-89b9-62629792efee" containerName="registry-server" containerID="cri-o://914d9d8d7a1697743b5bff33786c366f0996eb0ed9f378eb8569bd5c0a7d9812" gracePeriod=2 Nov 27 11:11:36 crc kubenswrapper[4806]: I1127 11:11:36.924463 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.029433 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88723dc1-3227-4511-89b9-62629792efee-utilities\") pod \"88723dc1-3227-4511-89b9-62629792efee\" (UID: \"88723dc1-3227-4511-89b9-62629792efee\") " Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.029778 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bttbm\" (UniqueName: \"kubernetes.io/projected/88723dc1-3227-4511-89b9-62629792efee-kube-api-access-bttbm\") pod \"88723dc1-3227-4511-89b9-62629792efee\" (UID: \"88723dc1-3227-4511-89b9-62629792efee\") " Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.030044 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88723dc1-3227-4511-89b9-62629792efee-catalog-content\") pod \"88723dc1-3227-4511-89b9-62629792efee\" (UID: \"88723dc1-3227-4511-89b9-62629792efee\") " Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.030448 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88723dc1-3227-4511-89b9-62629792efee-utilities" (OuterVolumeSpecName: "utilities") pod "88723dc1-3227-4511-89b9-62629792efee" (UID: "88723dc1-3227-4511-89b9-62629792efee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.030791 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88723dc1-3227-4511-89b9-62629792efee-utilities\") on node \"crc\" DevicePath \"\"" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.047050 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88723dc1-3227-4511-89b9-62629792efee-kube-api-access-bttbm" (OuterVolumeSpecName: "kube-api-access-bttbm") pod "88723dc1-3227-4511-89b9-62629792efee" (UID: "88723dc1-3227-4511-89b9-62629792efee"). InnerVolumeSpecName "kube-api-access-bttbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.133351 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bttbm\" (UniqueName: \"kubernetes.io/projected/88723dc1-3227-4511-89b9-62629792efee-kube-api-access-bttbm\") on node \"crc\" DevicePath \"\"" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.142782 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88723dc1-3227-4511-89b9-62629792efee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88723dc1-3227-4511-89b9-62629792efee" (UID: "88723dc1-3227-4511-89b9-62629792efee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.236048 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88723dc1-3227-4511-89b9-62629792efee-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.377765 4806 generic.go:334] "Generic (PLEG): container finished" podID="88723dc1-3227-4511-89b9-62629792efee" containerID="914d9d8d7a1697743b5bff33786c366f0996eb0ed9f378eb8569bd5c0a7d9812" exitCode=0 Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.377859 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66l69" event={"ID":"88723dc1-3227-4511-89b9-62629792efee","Type":"ContainerDied","Data":"914d9d8d7a1697743b5bff33786c366f0996eb0ed9f378eb8569bd5c0a7d9812"} Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.378337 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66l69" event={"ID":"88723dc1-3227-4511-89b9-62629792efee","Type":"ContainerDied","Data":"dbb2c5de440d50a7270bf32a68e9d1318d7c6ff1f88fd9e4e8415ad50f31bbaa"} Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.378379 4806 scope.go:117] "RemoveContainer" containerID="914d9d8d7a1697743b5bff33786c366f0996eb0ed9f378eb8569bd5c0a7d9812" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.377896 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66l69" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.414903 4806 scope.go:117] "RemoveContainer" containerID="24e14d0279dbb633f3aa65f40cc31f24609f9d33ddcb4241e0e1985dc334bd5a" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.463136 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-66l69"] Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.474217 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-66l69"] Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.492804 4806 scope.go:117] "RemoveContainer" containerID="8d61b2293d73b991af200ec07e333bc664beaf9a4e90c8ee66b2356ae8e98b82" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.519452 4806 scope.go:117] "RemoveContainer" containerID="914d9d8d7a1697743b5bff33786c366f0996eb0ed9f378eb8569bd5c0a7d9812" Nov 27 11:11:37 crc kubenswrapper[4806]: E1127 11:11:37.520151 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"914d9d8d7a1697743b5bff33786c366f0996eb0ed9f378eb8569bd5c0a7d9812\": container with ID starting with 914d9d8d7a1697743b5bff33786c366f0996eb0ed9f378eb8569bd5c0a7d9812 not found: ID does not exist" containerID="914d9d8d7a1697743b5bff33786c366f0996eb0ed9f378eb8569bd5c0a7d9812" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.520268 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"914d9d8d7a1697743b5bff33786c366f0996eb0ed9f378eb8569bd5c0a7d9812"} err="failed to get container status \"914d9d8d7a1697743b5bff33786c366f0996eb0ed9f378eb8569bd5c0a7d9812\": rpc error: code = NotFound desc = could not find container \"914d9d8d7a1697743b5bff33786c366f0996eb0ed9f378eb8569bd5c0a7d9812\": container with ID starting with 914d9d8d7a1697743b5bff33786c366f0996eb0ed9f378eb8569bd5c0a7d9812 not found: ID does not exist" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.520320 4806 scope.go:117] "RemoveContainer" containerID="24e14d0279dbb633f3aa65f40cc31f24609f9d33ddcb4241e0e1985dc334bd5a" Nov 27 11:11:37 crc kubenswrapper[4806]: E1127 11:11:37.520912 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24e14d0279dbb633f3aa65f40cc31f24609f9d33ddcb4241e0e1985dc334bd5a\": container with ID starting with 24e14d0279dbb633f3aa65f40cc31f24609f9d33ddcb4241e0e1985dc334bd5a not found: ID does not exist" containerID="24e14d0279dbb633f3aa65f40cc31f24609f9d33ddcb4241e0e1985dc334bd5a" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.521001 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e14d0279dbb633f3aa65f40cc31f24609f9d33ddcb4241e0e1985dc334bd5a"} err="failed to get container status \"24e14d0279dbb633f3aa65f40cc31f24609f9d33ddcb4241e0e1985dc334bd5a\": rpc error: code = NotFound desc = could not find container \"24e14d0279dbb633f3aa65f40cc31f24609f9d33ddcb4241e0e1985dc334bd5a\": container with ID starting with 24e14d0279dbb633f3aa65f40cc31f24609f9d33ddcb4241e0e1985dc334bd5a not found: ID does not exist" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.521032 4806 scope.go:117] "RemoveContainer" containerID="8d61b2293d73b991af200ec07e333bc664beaf9a4e90c8ee66b2356ae8e98b82" Nov 27 11:11:37 crc kubenswrapper[4806]: E1127 11:11:37.521627 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d61b2293d73b991af200ec07e333bc664beaf9a4e90c8ee66b2356ae8e98b82\": container with ID starting with 8d61b2293d73b991af200ec07e333bc664beaf9a4e90c8ee66b2356ae8e98b82 not found: ID does not exist" containerID="8d61b2293d73b991af200ec07e333bc664beaf9a4e90c8ee66b2356ae8e98b82" Nov 27 11:11:37 crc kubenswrapper[4806]: I1127 11:11:37.521697 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d61b2293d73b991af200ec07e333bc664beaf9a4e90c8ee66b2356ae8e98b82"} err="failed to get container status \"8d61b2293d73b991af200ec07e333bc664beaf9a4e90c8ee66b2356ae8e98b82\": rpc error: code = NotFound desc = could not find container \"8d61b2293d73b991af200ec07e333bc664beaf9a4e90c8ee66b2356ae8e98b82\": container with ID starting with 8d61b2293d73b991af200ec07e333bc664beaf9a4e90c8ee66b2356ae8e98b82 not found: ID does not exist" Nov 27 11:11:38 crc kubenswrapper[4806]: I1127 11:11:38.152735 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88723dc1-3227-4511-89b9-62629792efee" path="/var/lib/kubelet/pods/88723dc1-3227-4511-89b9-62629792efee/volumes" Nov 27 11:11:41 crc kubenswrapper[4806]: I1127 11:11:41.119309 4806 scope.go:117] "RemoveContainer" containerID="3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" Nov 27 11:11:41 crc kubenswrapper[4806]: E1127 11:11:41.120514 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:11:56 crc kubenswrapper[4806]: I1127 11:11:56.117888 4806 scope.go:117] "RemoveContainer" containerID="3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" Nov 27 11:11:56 crc kubenswrapper[4806]: E1127 11:11:56.122562 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" Nov 27 11:12:08 crc kubenswrapper[4806]: I1127 11:12:08.118586 4806 scope.go:117] "RemoveContainer" containerID="3083fff4eba8a2cd948f69d5b12052992d9c7d53f2ca513a0b296628ec2291b4" Nov 27 11:12:08 crc kubenswrapper[4806]: E1127 11:12:08.120328 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mq9p_openshift-machine-config-operator(4ea0cad1-8e8b-458c-a943-7191b36e35a3)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mq9p" podUID="4ea0cad1-8e8b-458c-a943-7191b36e35a3" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515112031022024430 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015112031023017346 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015112022360016475 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015112022360015445 5ustar corecore